scholarly journals Optimal Schedules for Parallelizing Anytime Algorithms: The Case of Shared Resources

2003 ◽  
Vol 19 ◽  
pp. 73-138 ◽  
Author(s):  
L. Finkelstein ◽  
S. Markovitch ◽  
E. Rivlin

The performance of anytime algorithms can be improved by simultaneously solving several instances of algorithm-problem pairs. These pairs may include different instances of a problem (such as starting from a different initial state), different algorithms (if several alternatives exist), or several runs of the same algorithm (for non-deterministic algorithms). In this paper we present a methodology for designing an optimal scheduling policy based on the statistical characteristics of the algorithms involved. We formally analyze the case where the processes share resources (a single-processor model), and provide an algorithm for optimal scheduling. We analyze, theoretically and empirically, the behavior of our scheduling algorithm for various distribution types. Finally, we present empirical results of applying our scheduling algorithm to the Latin Square problem.

Author(s):  
Erhun Özkan

A fork-join processing network is a queueing network in which tasks associated with a job can be processed simultaneously. Fork-join processing networks are prevalent in computer systems, healthcare, manufacturing, project management, justice systems, and so on. Unlike the conventional queueing networks, fork-join processing networks have synchronization constraints that arise because of the parallel processing of tasks and can cause significant job delays. We study scheduling in fork-join processing networks with multiple job types and parallel shared resources. Jobs arriving in the system fork into arbitrary number of tasks, then those tasks are processed in parallel, and then they join and leave the network. There are shared resources processing multiple job types. We study the scheduling problem for those shared resources (i.e., which type of job to prioritize at any given time) and propose an asymptotically optimal scheduling policy in diffusion scale.


2002 ◽  
Vol 51 (4) ◽  
pp. 444-448 ◽  
Author(s):  
Chan-Ik Park ◽  
Tae-Young Choe

Author(s):  
Ravi Mahadevan ◽  
Neelamegam Anbazhagan

<span>Online Nowadays, the enterprises &amp; individuals are contributing their workloads on cloud service providers which are going to increase on daily basis. There are   large amount CSP are available to offer virtualized and dynamic resource on pay and use basis. However, there are almost CSP failed to maintain quality of service (QOS) and minimal resource optimization. Some of the existing approaches are highly dedicated on scheduling policy but, it does not considered reliable services with optimized QOS. To offer best solution of above problem, the framework proposes Enhanced Minimal Resource Optimization based Scheduling Algorithm to minimize the resources and maintain the QOS.  The method avoids delay in Request-Response model in cloud environment. To avoid overload for resource allocation, the proposed design utilized optimized scheduling policy.  Proposed mechanisms utilized optimized service brokering policy to reduce the delay response in cloud environment. The framework also help cloud user to prefer best CSP according to their prior services. The method offers rising trend of resource based structure to reduce the placement churn extensively. Proposed system utilized efficient scheduling policy to transmit data request to CSP with minimal data processing time. The entire utilization is to improve the QOS of cloud service provider in the features of multi-dimensional resource. Based on experimental evaluations, proposed technique improves the CPT (Computation Processing Time) 301.72 milliseconds, BU (Bandwidth Utilization) 20 Mbps, CPUU (CPU Utilization) 5% &amp; MRU (Memory Resource Utilization) 3% on given input parameters compare than existing methodology.</span>


Author(s):  
Zhenyang Lei ◽  
Xiangdong Lei ◽  
Jun Long

Shared resources on the multicore chip, such as main memory, are increasingly becoming a point of contention. Traditional real-time task scheduling policies focus on solely on the CPU, and do not take in account memory access and cache effects. In this paper, we propose parallel real-time tasks scheduling (PRTTS) policy on multicore platforms. Each set of tasks is represented as a directed acyclic graph (DAG). The priorities of tasks are assigned according to task periods Rate Monotonic (RM). Each task is composed of three phases. The first phase is read memory stage, the second phase is execution phase and the third phase is write memory phase. The tasks use locks and critical sections to protect data access. The global scheduler maintains the task pool in which tasks are ready to be executed which can run on any core. PRTTS scheduling policy consists of two levels: the first level scheduling schedules ready real-time tasks in the task pool to cores, and the second level scheduling schedules real-time tasks on cores. Tasks can preempt the core on running tasks of low priority. The priorities of tasks which want to access memory are dynamically increased above all tasks that do not access memory. When the data accessed by a task is in the cache, the priority of the task is raised to the highest priority, and the task is scheduled immediately to preempt the core on running the task not accessing memory. After accessing memory, the priority of these tasks is restored to the original priority and these tasks are pended, the preempted task continues to run on the core. This paper analyzes the schedulability of PRTTS scheduling policy. We derive an upper-bound on the worst-case response-time for parallel real-time tasks. A series of extensive simulation experiments have been performed to evaluate the performance of proposed PRTTS scheduling policy. The results of simulation experiment show that PRTTS scheduling policy offers better performance in terms of core utilization and schedulability rate of tasks.


Sign in / Sign up

Export Citation Format

Share Document