scholarly journals Phase-Aware Cache Partitioning to Target Both Turnaround Time and System Performance

2020 ◽  
Vol 31 (11) ◽  
pp. 2556-2568 ◽  
Author(s):  
Lucia Pons ◽  
Julio Sahuquillo ◽  
Vicent Selfa ◽  
Salvador Petit ◽  
Julio Pons
2018 ◽  
Vol 7 (2) ◽  
pp. 837
Author(s):  
S Gokuldev ◽  
Jathin R

Performing scheduling of tasks with low energy consumption with high performance is one of the major concerns in distributed computing. Most of the existing systems have achieved improved energy efficiency but compromised with QoS metrics such as makespan and resource utilization. A resource scheduling strategy for wireless clusters is proposed by making careful considerations on decisions that would im-prove the battery life of nodes. The proposed strategy also incorporates monitoring system with in the clusters for optimizing the system performance as well as energy consumption. The system ensures “Any case zero loss" performance wherein each cluster will be monitored by at least one cluster monitor. This is implemented by using predictive calculation at each cluster monitor to communicate only if absolutely essential, during assigning jobs to resources, selecting optimal resources by assigning the jobs to the most power efficient resource among the available idle resources within the cluster. The experimental result ensures improved system performance with low power consumption in homogeneous computing environment. The resource sharing strategy is experimentally analyzed, considering the important performance metrics such as starvation deadline, turnaround time, miss hit count through simulations. Significant results were observed with improved efficiency.  


2010 ◽  
Vol 439-440 ◽  
pp. 1587-1594
Author(s):  
Shuo Li ◽  
Feng Wu

In a chip-multiprocessor with a shared cache structure , the competing accesses from different applications degrade the system performance.The accesses degrade the performance and result in non-predicting executing time. Cache partitioning techniques can exclusively partition the shared cache among multiple competing applications. In this paper, the authors design the framework of Process priority-based Multithread Cache Partitioning(PP-MCP),a dynamic shared cache partitioning mechanism to improve the performance of multi-threaded multi-programmed workloads. The framework includes a miss rate monitor , called Application-oriented Miss Rate Monitor (AMRM) , which dynamically collect s miss rate information of multiple multi-threaded applications on different cache partitions , and process priority-based weighted cache partitioning algorithm ,which extends traditional miss rate oriented cache partition algorithms.The algorithm allocates Cache in sequence of the value of the process priority and it ensures that the highest priority process will get enough Cache space; and the applications with more threads tend to get more shared cache in order to improve t he overall system performance. Experiments show that PP-MCP has better IPC throughput and weighted speedup. Specifically , for multi-threaded multi-programmed scientific computing workloads , PP-MCP-1 improves throughput by up to 20% and on average 10 % over PP-MCP-0.


2017 ◽  
Vol 7 (2) ◽  
pp. 10-26 ◽  
Author(s):  
Saad Bani-Mohammad

Contiguous processor allocation is useful for security and accounting reasons. This is due to the allocated jobs are separated from one another, where each sub-mesh of processors is allocated to an exclusive job request, and the allocated sub-mesh has the same size and shape of the requested job. The size and shape constraint leads to high processor fragmentation. Most recent contiguous allocation strategies suggested for 3D mesh-connected multiconputers try all possible orientations of an allocation request when allocation fails for the requested orientation, which reduces processor fragmentation and hence improves system performance. However, none of them considers all shapes of the request in the process of allocation. To generalize this restricted rotation, we propose, in this paper, a new contiguous allocation strategy for 3D mesh-connected multicomputers, referred to as All Shapes Busy List (ASBL for short), which takes into consideration all possible contiguous request shapes when attempting allocation for a job request. ASBL depends on the list of allocated sub-meshes, in the method suggested in (Bani-Mohammad et al., 2006), for selecting an allocated sub-mesh. The performance of the proposed ASBL allocation strategy has been evaluated considering several important scheduling strategies under a variety of system loads based on different job size distributions. The simulation results have shown that the ASBL allocation strategy improves system performance in terms of parameters such as the average turnaround time of jobs and system utilization under all scheduling strategies considered.


2010 ◽  
Vol 439-440 ◽  
pp. 1223-1229
Author(s):  
Shuo Li ◽  
Gao Chao Xu ◽  
Yu Shuang Dong ◽  
Feng Wu

With the development of microelectronics technology, Chip Multi-Processor (CMP) or multi-core design has become a mainstream choice for major microprocessor vendors. But in a chip-multiprocessor with a shared cache structure , the competing accesses from different applications degrade the system performance , resulting in non-optimal performance and non-predicting executing time. Cache partitioning techniques can exclusively partition the shared cache among multiple competing applications. In this paper, we first introduce the problems caused by Cache pollution in multicore processor structure; then present the different methods of Cache partitioning in multicore processor structure¬ --categorizing them based on the different metrics. And finally, we discuss some possible directions for future research in the area.


Author(s):  
Kuo-Chan Huang ◽  
Po-Chi Shih ◽  
Yeh-Ching Chung

Most current grid environments are established through collaboration among a group of participating sites which volunteer to provide free computing resources. Therefore, feasible load sharing policies that benefit all sites are an important incentive for attracting computing sites to join and stay in a grid environment. Moreover, a grid environment is usually heterogeneous in nature at least for different computing speeds at different participating sites. This chapter explores the feasibility and effectiveness of load sharing activities in a heterogeneous computational grid. Several issues are discussed including site selection policies as well as feasible load sharing mechanisms. Promising policies are evaluated in a series of simulations based on workloads derived from real traces. The results show that grid computing is capable of significantly improving the overall system performance in terms of average turnaround time for user jobs.


2004 ◽  
Vol 13 (02) ◽  
pp. 159-181 ◽  
Author(s):  
JIN-WOOK BAEK ◽  
JAE-HEUNG YEO ◽  
GYU-TAE KIM ◽  
HEON-YOUNG YEOM

Two significant performance factors in Mobile Agent Planning (MAP) for distributed information retrieval are the number of mobile agents and the total execution time. Using fewer mobile agents results in less network traffic and consumes less bandwidth. Regardless of the number of agents used, the total execution time for a task must be kept to a minimum. A retrieval service must minimize both these factors for better system performance, and at the same time, it must be able to supply the required information to users as quickly as possible. In this paper, we propose heuristic algorithms, called Cost-Effective MAP (CEMAP), to minimize both the number of mobile agents and the total execution time under the condition that the turnaround time is kept to a minimum. Although these algorithms tend to slightly increase the planning cost, a simulation study shows that these algorithms enhance the system performance significantly. By adopting these algorithms, systems can maintain lower network traffic while satisfying the minimum turnaround time.


Author(s):  
P. B. Basham ◽  
H. L. Tsai

The use of transmission electron microscopy (TEM) to support process development of advanced microelectronic devices is often challenged by a large amount of samples submitted from wafer fabrication areas and specific-spot analysis. Improving the TEM sample preparation techniques for a fast turnaround time is critical in order to provide a timely support for customers and improve the utilization of TEM. For the specific-area sample preparation, a technique which can be easily prepared with the least amount of effort is preferred. For these reasons, we have developed several techniques which have greatly facilitated the TEM sample preparation.For specific-area analysis, the use of a copper grid with a small hole is found to be very useful. With this small-hole grid technique, TEM sample preparation can be proceeded by well-established conventional methods. The sample is first polished to the area of interest, which is then carefully positioned inside the hole. This polished side is placed against the grid by epoxy Fig. 1 is an optical image of a TEM cross-section after dimpling to light transmission.


1960 ◽  
Author(s):  
S. Seidenstein ◽  
R. Chernikoff ◽  
F. V. Taylor

Author(s):  
Christopher Wickens ◽  
Jack Isreal ◽  
Gregory McCarthy ◽  
Daniel Gopher ◽  
Emanuel Donchin

Sign in / Sign up

Export Citation Format

Share Document