Network and data location aware approach for simultaneous job scheduling and data replication in large-scale data grid environments

2014 ◽  
Vol 8 (3) ◽  
pp. 391-408 ◽  
Author(s):  
Najme Mansouri
Author(s):  
Abdenour Lazeb ◽  
Riad Mokadem ◽  
Ghalem Belalem

Applications produce huge volumes of data that are distributed on remote and heterogeneous sites. This generates problems related to access and sharing data. As a result, managing data in large-scale environments is a real challenge. In this context, large-scale data management systems often use data replication, a well-known technique that treats generated problems by storing multiple copies of data, called replicas, across multiple nodes. Most of the replication strategies in these environments are difficult to adapt to cloud environments. They aim to achieve the best performance of the system without meeting the important objectives of the cloud provider. This article proposes a new dynamic replication strategy. The proposed algorithm significantly improves provider gain without neglecting customer satisfaction.


Author(s):  
Zhenghua Xue ◽  
Jianhui Li ◽  
Yuanchun Zhou ◽  
Yang Zhang ◽  
Geng Shen

2015 ◽  
Vol 4 (1) ◽  
pp. 163 ◽  
Author(s):  
Alireza Saleh ◽  
Reza Javidan ◽  
Mohammad Taghi FatehiKhajeh

<p>Nowadays, scientific applications generate a huge amount of data in terabytes or petabytes. Data grids currently proposed solutions to large scale data management problems including efficient file transfer and replication. Data is typically replicated in a Data Grid to improve the job response time and data availability. A reasonable number and right locations for replicas has become a challenge in the Data Grid. In this paper, a four-phase dynamic data replication algorithm based on Temporal and Geographical locality is proposed. It includes: 1) evaluating and identifying the popular data and triggering a replication operation when the popularity data passes a dynamic threshold; 2) analyzing and modeling the relationship between system availability and the number of replicas, and calculating a suitable number of new replicas; 3) evaluating and identifying the popular data in each site, and placing replicas among them; 4) removing files with least cost of average access time when encountering insufficient space for replication. The algorithm was tested using a grid simulator, OptorSim developed by European Data Grid Projects. The simulation results show that the proposed algorithm has better performance in comparison with other algorithms in terms of job execution time, effective network usage and percentage of storage filled.</p>


2008 ◽  
Vol 4 (2) ◽  
pp. 103-110
Author(s):  
Mohammed Radi ◽  
Ali Mamat ◽  
M. Mat Deris ◽  
Hamidah Ibrahim ◽  
Subramaniam Shamala

2010 ◽  
Vol 26 (1) ◽  
pp. 12-20 ◽  
Author(s):  
José M. Pérez ◽  
Félix García-Carballeira ◽  
Jesús Carretero ◽  
Alejandro Calderón ◽  
Javier Fernández

Author(s):  
Mohammed Radi ◽  
Ali Mamat ◽  
M.Mat Deris ◽  
Hamidah Ibrahim ◽  
Subramaniam Shamala

2009 ◽  
Vol 28 (11) ◽  
pp. 2737-2740
Author(s):  
Xiao ZHANG ◽  
Shan WANG ◽  
Na LIAN

Sign in / Sign up

Export Citation Format

Share Document