Efficient Big Data Transfer Using Bandwidth Reservation Service in High-Performance Networks

Author(s):  
Liudong Zuo ◽  
Michelle Mengxia Zhu
2019 ◽  
Vol E102.D (8) ◽  
pp. 1478-1488
Author(s):  
Eun-Sung JUNG ◽  
Si LIU ◽  
Rajkumar KETTIMUTHU ◽  
Sungwook CHUNG

Author(s):  
Daqing Yun ◽  
Chase Q. Wu

High-performance networks featuring advance bandwidth reservation have been developed and deployed to support big data transfer in extreme-scale scientific applications. The performance of such big data transfer largely depends on the transport protocols being used. For a given protocol in a given network environment, different parameter settings may lead to different performance, and oftentimes the default settings do not yield the best performance. It is, however, impractical to conduct an exhaustive search in the large parameter space of transport protocols for a set of suitable parameter values. This chapter proposes a stochastic approximation-based transport profiler, namely FastProf, to quickly determine the optimal operational zone of a protocol over dedicated connections. The proposed method is evaluated using both emulations based on real-life measurements and experiments over physical connections. The results show that FastProf significantly reduces the profiling overhead while achieving a comparable level of transport performance with the exhaustive search-based approach.


2019 ◽  
Vol 6 (1) ◽  
Author(s):  
Mahdi Torabzadehkashi ◽  
Siavash Rezaei ◽  
Ali HeydariGorji ◽  
Hosein Bobarshad ◽  
Vladimir Alves ◽  
...  

AbstractIn the era of big data applications, the demand for more sophisticated data centers and high-performance data processing mechanisms is increasing drastically. Data are originally stored in storage systems. To process data, application servers need to fetch them from storage devices, which imposes the cost of moving data to the system. This cost has a direct relation with the distance of processing engines from the data. This is the key motivation for the emergence of distributed processing platforms such as Hadoop, which move process closer to data. Computational storage devices (CSDs) push the “move process to data” paradigm to its ultimate boundaries by deploying embedded processing engines inside storage devices to process data. In this paper, we introduce Catalina, an efficient and flexible computational storage platform, that provides a seamless environment to process data in-place. Catalina is the first CSD equipped with a dedicated application processor running a full-fledged operating system that provides filesystem-level data access for the applications. Thus, a vast spectrum of applications can be ported for running on Catalina CSDs. Due to these unique features, to the best of our knowledge, Catalina CSD is the only in-storage processing platform that can be seamlessly deployed in clusters to run distributed applications such as Hadoop MapReduce and HPC applications in-place without any modifications on the underlying distributed processing framework. For the proof of concept, we build a fully functional Catalina prototype and a CSD-equipped platform using 16 Catalina CSDs to run Intel HiBench Hadoop and HPC benchmarks to investigate the benefits of deploying Catalina CSDs in the distributed processing environments. The experimental results show up to 2.2× improvement in performance and 4.3× reduction in energy consumption, respectively, for running Hadoop MapReduce benchmarks. Additionally, thanks to the Neon SIMD engines, the performance and energy efficiency of DFT algorithms are improved up to 5.4× and 8.9×, respectively.


Sign in / Sign up

Export Citation Format

Share Document