scholarly journals Near real-time streaming analysis of big fusion data

Author(s):  
Ralph Kube ◽  
Randy Michael Churchill ◽  
Choong Seock Chang ◽  
Jong Choi ◽  
Ruonan Wang ◽  
...  

Abstract Experiments on fusion plasmas produce high-dimensional data time series with ever increasing magnitude and velocity, but turn-around times for analysis of this data have not kept up. For example, many data analysis tasks are often performed in a manual, ad-hoc manner some time after an experiment. In this article we introduce the DELTA framework that facilitates near real-time streaming analysis of big and fast fusion data. By streaming measurement data from fusion experiments to a high-performance compute center, DELTA allows computationally expensive data analysis tasks to be performed in between plasma pulses. This article describe the modular and expandable software architecture of DELTA and present performance benchmarks of individual components as well as of an example workflows. Focusing on a streaming analysis workflow where Electron cyclotron emission imaging (ECEi) data measured at KSTAR on NERSC's supercomputer we routinely observe data transfer rates of about 4 Gigabit per second. At NERSC, a demanding turbulence analysis workflow effectively utilizes multiple nodes and graphical processing units and executes in under 5 minutes. We further discuss how DELTA uses modern database systems and container orchestration services to provide web-based real-time data visualization. For the case of ECEi data we demonstrate how data visualizations can be augmented with outputs from machine learning models. By providing session leaders and physics operators results of higher order data analysis using live visualizations may make more informed decisions on how to configure the machine for the next shot.

2021 ◽  
Vol 14 (10) ◽  
pp. 1818-1831
Author(s):  
Rudi Poepsel-Lemaitre ◽  
Martin Kiefer ◽  
Joscha von Hein ◽  
Jorge-Arnulfo Quiané-Ruiz ◽  
Volker Markl

In pursuit of real-time data analysis, approximate summarization structures, i.e., synopses, have gained importance over the years. However, existing stream processing systems, such as Flink, Spark, and Storm, do not support synopses as first class citizens, i.e., as pipeline operators. Synopses' implementation is upon users. This is mainly because of the diversity of synopses, which makes a unified implementation difficult. We present Condor, a framework that supports synopses as first class citizens. Condor facilitates the specification and processing of synopsis-based streaming jobs while hiding all internal processing details. Condor's key component is its model that represents synopses as a particular case of windowed aggregate functions. An inherent divide and conquer strategy allows Condor to efficiently distribute the computation, allowing for high-performance and linear scalability. Our evaluation shows that Condor outperforms existing approaches by up to a factor of 75x and that it scales linearly with the number of cores.


2018 ◽  
Vol 19 (S18) ◽  
Author(s):  
Ahmed Sanaullah ◽  
Chen Yang ◽  
Yuri Alexeev ◽  
Kazutomo Yoshii ◽  
Martin C. Herbordt

2019 ◽  
Vol 26 (1) ◽  
pp. 244-252 ◽  
Author(s):  
Shibom Basu ◽  
Jakub W. Kaminski ◽  
Ezequiel Panepucci ◽  
Chia-Ying Huang ◽  
Rangana Warshamanage ◽  
...  

At the Swiss Light Source macromolecular crystallography (MX) beamlines the collection of serial synchrotron crystallography (SSX) diffraction data is facilitated by the recent DA+ data acquisition and analysis software developments. The SSX suite allows easy, efficient and high-throughput measurements on a large number of crystals. The fast continuous diffraction-based two-dimensional grid scan method allows initial location of microcrystals. The CY+ GUI utility enables efficient assessment of a grid scan's analysis output and subsequent collection of multiple wedges of data (so-called minisets) from automatically selected positions in a serial and automated way. The automated data processing (adp) routines adapted to the SSX data collection mode provide near real time analysis for data in both CBF and HDF5 formats. The automatic data merging (adm) is the latest extension of the DA+ data analysis software routines. It utilizes the sxdm (SSX data merging) package, which provides automatic online scaling and merging of minisets and allows identification of a minisets subset resulting in the best quality of the final merged data. The results of both adp and adm are sent to the MX MongoDB database and displayed in the web-based tracker, which provides the user with on-the-fly feedback about the experiment.


Author(s):  
Rashima Mahajan ◽  
Pragya Gupta

The progressive research in the field of internet of things provides a platform to develop high performance and robust automated systems to control external devices via internet data transfer and cloud computing. The present emerging IoT research including user-friendly and easily-wearable sensors and signal acquisition techniques have made it possible to expand the IoT application areas towards healthcare sector. This chapter aims at providing a rationale behind development of IoT applications in healthcare, architecture details of internet of healthcare things (IoHT), and highlights a step-by-step development of IoT-based heart rate measurement and monitoring system using Arduino. The developed module has been advanced to transmit data over the internet on the ThingSpeak channel to allow remote monitoring in real time. This may help to improve/restore useful life among cardiac patients via real-time monitoring through remote locations.


Sign in / Sign up

Export Citation Format

Share Document