Scientific Data Processing Using Mapreduce in Cloud Environments

2013 ◽  
Vol 12 (23) ◽  
pp. 7869-7873
Author(s):  
Kong Xiangsheng
Author(s):  
Y. Xu ◽  
L. P. Xin ◽  
X. H. Han ◽  
H. B. Cai ◽  
L. Huang ◽  
...  

GWAC will have been built an integrated FOV of 5,000 degree2 and have already built 1,800 square degree2. The limit magnitude of a 10-second exposure image in the moonless night is 16R. In each observation night, GWAC produces about 0.7TB of raw data, and the data processing pipeline generates millions of single frame alerts. We describe the GWAC Data Processing and Management System (GPMS), including hardware architecture, database, detection-filtering-validation of transient candidates, data archiving, and user interfaces for the check of transient and the monitor of the system. GPMS combines general technology and software in astronomy and computer field, and use some advanced technologies such as deep learning. Practical results show that GPMS can fully meet the scientific data processing requirement of GWAC. It can online accomplish the detection, filtering and validation of millions of transient candidates, and feedback the final results to the astronomer in real-time. During the observation from October of 2018 to December of 2019, we have already found 102 transients.


2014 ◽  
Vol 60 ◽  
pp. 241-249 ◽  
Author(s):  
Krista Gaustad ◽  
Tim Shippert ◽  
Brian Ermold ◽  
Sherman Beus ◽  
Jeff Daily ◽  
...  

2009 ◽  
Vol 5 (S261) ◽  
pp. 296-305 ◽  
Author(s):  
Lennart Lindegren

AbstractThe scientific objectives of the Gaia mission cover areas of galactic structure and evolution, stellar astrophysics, exoplanets, solar system physics, and fundamental physics. Astrometrically, its main contribution will be the determination of millions of absolute stellar parallaxes and the establishment of a very accurate, dense and faint non-rotating optical reference frame. With a planned launch in spring 2012, the project is in its advanced implementation phase. In parallel, preparations for the scientific data processing are well under way within the Gaia Data Processing and Analysis Consortium. Final mission results are expected around 2021, but early releases of preliminary data are expected. This review summarizes the main science goals and overall organisation of the project, the measurement principle and core astrometric solution, and provide an updated overview of the expected astrometric performance.


2012 ◽  
Vol 24 (5) ◽  
pp. 327-334 ◽  
Author(s):  
Luis David Patiño-Lopez ◽  
Klaas Decanniere ◽  
Jose Antonio Gavira ◽  
Dominique Maes ◽  
Fermín Otalora

2021 ◽  
Vol 11 (13) ◽  
pp. 6200
Author(s):  
Jin-young Choi ◽  
Minkyoung Cho ◽  
Jik-Soo Kim

Recently, “Big Data” platform technologies have become crucial for distributed processing of diverse unstructured or semi-structured data as the amount of data generated increases rapidly. In order to effectively manage these Big Data, Cloud Computing has been playing an important role by providing scalable data storage and computing resources for competitive and economical Big Data processing. Accordingly, server virtualization technologies that are the cornerstone of Cloud Computing have attracted a lot of research interests. However, conventional hypervisor-based virtualization can cause performance degradation problems due to its heavily loaded guest operating systems and rigid resource allocations. On the other hand, container-based virtualization technology can provide the same level of service faster with a lightweight capacity by effectively eliminating the guest OS layers. In addition, container-based virtualization enables efficient cloud resource management by dynamically adjusting the allocated computing resources (e.g., CPU and memory) during the runtime through “Vertical Elasticity”. In this paper, we present our practice and experience of employing an adaptive resource utilization scheme for Big Data workloads in container-based cloud environments by leveraging the vertical elasticity of Docker, a representative container-based virtualization technique. We perform extensive experiments running several Big Data workloads on representative Big Data platforms: Apache Hadoop and Spark. During the workload executions, our adaptive resource utilization scheme periodically monitors the resource usage patterns of running containers and dynamically adjusts allocated computing resources that could result in substantial improvements in the overall system throughput.


1963 ◽  
Vol 55 (4) ◽  
pp. 29-32 ◽  
Author(s):  
P. A. D. de Maine ◽  
R. D. Seawright

Sign in / Sign up

Export Citation Format

Share Document