Performance evaluation of machine learning based big data processing framework for prediction of heart disease

Author(s):  
Abderrahmane Ed-Daoudy ◽  
Khalil Maalmi
2017 ◽  
Vol 102 (3) ◽  
pp. 2099-2116 ◽  
Author(s):  
Gunasekaran Manogaran ◽  
V. Vijayakumar ◽  
R. Varatharajan ◽  
Priyan Malarvizhi Kumar ◽  
Revathi Sundarasekar ◽  
...  

2019 ◽  
Vol 8 (9) ◽  
pp. 387 ◽  
Author(s):  
Silvino Pedro Cumbane ◽  
Gyozo Gidófalvi

Natural hazards result in devastating losses in human life, environmental assets and personal, and regional and national economies. The availability of different big data such as satellite imageries, Global Positioning System (GPS) traces, mobile Call Detail Records (CDRs), social media posts, etc., in conjunction with advances in data analytic techniques (e.g., data mining and big data processing, machine learning and deep learning) can facilitate the extraction of geospatial information that is critical for rapid and effective disaster response. However, disaster response systems development usually requires the integration of data from different sources (streaming data sources and data sources at rest) with different characteristics and types, which consequently have different processing needs. Deciding which processing framework to use for a specific big data to perform a given task is usually a challenge for researchers from the disaster management field. Therefore, this paper contributes in four aspects. Firstly, potential big data sources are described and characterized. Secondly, the big data processing frameworks are characterized and grouped based on the sources of data they handle. Then, a short description of each big data processing framework is provided and a comparison of processing frameworks in each group is carried out considering the main aspects such as computing cluster architecture, data flow, data processing model, fault-tolerance, scalability, latency, back-pressure mechanism, programming languages, and support for machine learning libraries, which are related to specific processing needs. Finally, a link between big data and processing frameworks is established, based on the processing provisioning for essential tasks in the response phase of disaster management.


Author(s):  
Snigdha Sen ◽  
Sonali Agarwal ◽  
Pavan Chakraborty ◽  
Krishna Pratap Singh

Procedia CIRP ◽  
2019 ◽  
Vol 83 ◽  
pp. 661-664 ◽  
Author(s):  
Yinghao Ye ◽  
Meilin Wang ◽  
Shuhong Yao ◽  
Jarvis N. Jiang ◽  
Qing Liu

2020 ◽  
Vol 10 (14) ◽  
pp. 4901
Author(s):  
Waleed Albattah ◽  
Rehan Ullah Khan ◽  
Khalil Khan

Processing big data requires serious computing resources. Because of this challenge, big data processing is an issue not only for algorithms but also for computing resources. This article analyzes a large amount of data from different points of view. One perspective is the processing of reduced collections of big data with less computing resources. Therefore, the study analyzed 40 GB data to test various strategies to reduce data processing. Thus, the goal is to reduce this data, but not to compromise on the detection and model learning in machine learning. Several alternatives were analyzed, and it is found that in many cases and types of settings, data can be reduced to some extent without compromising detection efficiency. Tests of 200 attributes showed that with a performance loss of only 4%, more than 80% of the data could be ignored. The results found in the study, thus provide useful insights into large data analytics.


Sign in / Sign up

Export Citation Format

Share Document