COMET: Client-Oriented METadata Service for Highly Available Distributed File Systems

Author(s):  
Ruini Xue ◽  
Lixiang Ao ◽  
Zhongyang Guan
2014 ◽  
Vol 2014 ◽  
pp. 1-8
Author(s):  
Jianwei Liao ◽  
Guoqiang Xiao ◽  
Xiaoning Peng

This paper presents a novel metadata management mechanism on the metadata server (MDS) for parallel and distributed file systems. In this technique, the client file system backs up the sent metadata requests, which have been handled by the metadata server, so that the MDS does not need to log metadata changes to nonvolatile storage for achieving highly available metadata service, as well as better performance improvement in metadata processing. As the client file system backs up certain sent metadata requests in its memory, the overhead for handling these backup requests is much smaller than that brought by the metadata server, while it adopts logging or journaling to yield highly available metadata service. The experimental results show that this newly proposed mechanism can significantly improve the speed of metadata processing and render a better I/O data throughput, in contrast to conventional metadata management schemes, that is, logging or journaling on MDS. Besides, a complete metadata recovery can be achieved by replaying the backup logs cached by all involved clients, when the metadata server has crashed or gone into nonoperational state exceptionally.


2020 ◽  
Vol 31 (2) ◽  
pp. 374-392 ◽  
Author(s):  
Jiang Zhou ◽  
Yong Chen ◽  
Weiping Wang ◽  
Shuibing He ◽  
Dan Meng

2021 ◽  
pp. 87-100
Author(s):  
Chuangwei Lin ◽  
Bowen Liu ◽  
Wei Zhou ◽  
Yueyue Xu ◽  
Xuyun Zhang ◽  
...  

2018 ◽  
Vol 29 (12) ◽  
pp. 2641-2657 ◽  
Author(s):  
Siyang Li ◽  
Fenlin Liu ◽  
Jiwu Shu ◽  
Youyou Lu ◽  
Tao Li ◽  
...  

Author(s):  
Sai Wu ◽  
Gang Chen ◽  
Xianke Zhou ◽  
Zhenjie Zhang ◽  
Anthony K. H. Tung ◽  
...  

2013 ◽  
Vol 49 (6) ◽  
pp. 2645-2652 ◽  
Author(s):  
Zhipeng Tan ◽  
Wei Zhou ◽  
Dan Feng ◽  
Wenhua Zhang

Electronics ◽  
2021 ◽  
Vol 10 (12) ◽  
pp. 1471
Author(s):  
Jun-Yeong Lee ◽  
Moon-Hyun Kim ◽  
Syed Asif Raza Raza Shah ◽  
Sang-Un Ahn ◽  
Heejun Yoon ◽  
...  

Data are important and ever growing in data-intensive scientific environments. Such research data growth requires data storage systems that play pivotal roles in data management and analysis for scientific discoveries. Redundant Array of Independent Disks (RAID), a well-known storage technology combining multiple disks into a single large logical volume, has been widely used for the purpose of data redundancy and performance improvement. However, this requires RAID-capable hardware or software to build up a RAID-enabled disk array. In addition, it is difficult to scale up the RAID-based storage. In order to mitigate such a problem, many distributed file systems have been developed and are being actively used in various environments, especially in data-intensive computing facilities, where a tremendous amount of data have to be handled. In this study, we investigated and benchmarked various distributed file systems, such as Ceph, GlusterFS, Lustre and EOS for data-intensive environments. In our experiment, we configured the distributed file systems under a Reliable Array of Independent Nodes (RAIN) structure and a Filesystem in Userspace (FUSE) environment. Our results identify the characteristics of each file system that affect the read and write performance depending on the features of data, which have to be considered in data-intensive computing environments.


Sign in / Sign up

Export Citation Format

Share Document