IMPROVING DATA AVAILABILITY USING HYBRID REPLICATION TECHNIQUE IN PEER-TO-PEER ENVIRONMENTS

2004 ◽  
Vol 05 (03) ◽  
pp. 299-312 ◽  
Author(s):  
M. D. MUSTAFA ◽  
B. NATHRAH ◽  
M. H. SUZURI ◽  
M. T. ABU OSMAN

Replication is an important technique in peer-to-peer environment, where it increases data availability and accessibility to users despite site or communication failure. However, determining the number of replication and where to replicate the data are the major issues. This paper proposes a hybrid replication model for fixed and mobile network in order to achieve high data availability. For the fixed network, a data will be replicated synchronously in a diagonal manner of logical grid structure, while for the mobile network, a data will be replicated asynchronously based on commonly visited sites for each user. In comparison to the previous techniques, diagonal replication technique (DRG) on fixed network requires lower communication cost for an operation, while providing higher data availability, which is preferred for large systems.

2001 ◽  
Vol 02 (03) ◽  
pp. 317-329 ◽  
Author(s):  
MUSTAFA MAT DERIS ◽  
ALI MAMAT ◽  
PUA CHAI SENG ◽  
MOHD YAZID SAMAN

This article addresses the performance of data replication protocol in terms of data availability and communication costs. Specifically, we present a new protocol called Three Dimensional Grid Structure (TDGS) protocol, to manage data replication in distributed system. The protocol provides high availability for read and write operations with limited fault-tolerance at low communication cost. With TDGS protocol, a read operation is limited to two data copies, while a write operation is required with minimal number of copies. In comparison to other protocols. TDGS requires lower communication cost for an operation, while providing higher data availability.


Author(s):  
Zulaile Mabni ◽  
Rohaya Latip ◽  
Hamidah Ibrahim ◽  
Azizol Abdullah

Data replication is widely used to provide high data availability, and increase the performance of the distributed systems. Many replica control protocols have been proposed in distributed and grid environments that achieved both high performance and availability. However, the previously proposed protocols still require a bigger number of replicas for read and write operations which are not suitable for a large scale system such as data grid. In this paper, a new replica control protocol called Clusteringbased Hybrid (CBH) has been proposed for managing the data in grid environments. We analyzed the communication cost and data availability for the operations and compared CBH protocol with recently proposed replica control protocols called Dynamic Hybrid (DH) protocol and Diagonal Replication in 2D Mesh (DR2M) protocol. To evaluate CBH protocol, a simulation model was implemented using Java. Our results show that for the read operations, CBH protocol improves the performance of communication cost and data availability compared to the DH and DR2M protocols.  


Author(s):  
Ellen Yu ◽  
Aparna Bhaskaran ◽  
Shang-Lin Chen ◽  
Zachary E. Ross ◽  
Egill Hauksson ◽  
...  

Abstract The Southern California Earthquake Data Center is hosting its earthquake catalog and seismic waveform archive in the Amazon Web Services (AWS) Open Dataset Program (s3://scedc-pds; us-west-2 region). The cloud dataset’s high data availability and scalability facilitate research that uses large volumes of data and computationally intensive processing. We describe the data archive and our rationale for the formats and data organization. We provide two simple examples to show how storing the data in AWS Simple Storage Service can benefit the analysis of large datasets. We share usage statistics of our data during the first year in the AWS Open Dataset Program. We also discuss the challenges and opportunities of a cloud-hosted archive.


2021 ◽  
Author(s):  
Hoda R.K. Nejad

With the emergence of wireless devices, service delivery for ad-hoc networks has started to attract a lot of attention recently. Ad-hoc networks provide an attractive solution for networking in the situations where network infrastructure or service subscription is not available. We believe that overlay networks, particularly peer-to-peer (P2P) systems, is a good abstraction for application design and deployment over ad-hoc networks. The principal benefit of this approach is that application states are only maintained by the nodes involved in the application execution and all other nodes only perform networking related functions. On the other hand, data access applications in Ad-hoc networks suffer from restricted resources. In this thesis, we explore how to use Cooperative Caching to improve data access efficiency in Ad-hoc network. We propose a Resource-Aware Cooperative Caching P2P system (RACC) for data access applications in Ad-hoc networks. The objective is to improve data availability by considering energy of each node, demand and supply of network. We evaluated and compared the performance of RACC with Simple Cache, CachePath and CacheData schemes. Our simulation results show that RACC improves the lay of query as well as energy usage of the network as compared to Simple Cache, CachePath and CacheData.


Author(s):  
Enrico Franchi ◽  
Michele Tomaiuolo

Social networking sites have deeply changed the perception of the web in the last years. Although the current approach to build social networking systems is to create huge centralized systems owned by a single company, such strategy has many drawbacks, e.g., lack of privacy, lack of anonymity, risks of censorship and operating costs. These issues contrast with some of the main requirements of information systems, including: (i) confidentiality, i.e., the interactions between a user and the system must remain private unless explicitly public; (ii) integrity; (iii) accountability; (iv) availability; (v) identity and anonymity. Moreover, social networking platforms are vulnerable to many kind of attacks: (i) masquerading, which occurs when a user disguises his identity and pretends to be another user; (ii) unauthorized access; (iii) denial of service; (iv) repudiation, which occurs when a user participates in an activity and later claims he did not; (v) eavesdropping; (vi) alteration of data; (vii) copy and replay attacks; and, in general, (viii) attacks making use of social engineering techniques. In order to overcome both the intrinsic defects of centralized systems and the general vulnerabilities of social networking platforms, many different approaches have been proposed, both as federated (i.e., consisting of multiple entities cooperating to provide the service, but usually distinct from users) or peer-to-peer systems (with users directly cooperating to provide the service); in this work the most interesting ones were reviewed. Eventually, the authors present their own approach to create a solid distributed social networking platform consisting in a novel peer-to-peer system that leverages existing, widespread and stable technologies such as distributed hash tables and BitTorrent. The topics considered in detail are: (i) anonymity and resilience to censorship; (ii) authenticatable contents; (iii) semantic interoperability using activity streams and weak semantic data formats for contacts and profiles; and (iv) data availability.


2013 ◽  
Vol 5 (2) ◽  
pp. 54-71 ◽  
Author(s):  
Jyh-Biau Chang ◽  
Po-Cheng Chen ◽  
Ce-Kuen Shieh ◽  
Jia-Hao Yang ◽  
Sheng-Hung Hsieh

Efficient information sharing is difficult to achieve in the scenario of emergency and rescue operations because there is no communication infrastructure at the disaster sites. In general, the network condition is relatively reliable in the intra-site environment but relatively unreliable in the inter-site environment. The network partitioning problem may occur between two sites. Although one can exploit the replication technique used in data grid to improve the information availability in emergency and rescue applications, the data consistency problem occurs between replicas. In this paper, the authors propose a middleware called “Seagull” to transparently manage the data availability and consistency issues of emergency and rescue applications. Seagull adopts the optimistic replication scheme to provide the higher data availability in the inter-site environment. It also adopts the pessimistic replication scheme to provide the stronger data consistency guarantee in the intra-site environment. Moreover, it adopts an adaptive consistency granularity strategy that achieves the better performance of the consistency management because this strategy provides the higher parallelism when the false sharing happens. Lastly, Seagull adopts the transparency data consistency management scheme, and thus the users do not need to modify their source codes to run on the Seagull.


Sign in / Sign up

Export Citation Format

Share Document