scholarly journals Research on Library Big Data Cleaning System based on Big Data Decision Analysis Needs

Author(s):  
Jianfeng Liao ◽  
Jianping You ◽  
Qun Zhang
2018 ◽  
Vol 7 (3.1) ◽  
pp. 63 ◽  
Author(s):  
R Revathy ◽  
R Aroul Canessane

Data are vital to help decision making. On the off chance that data have low veracity, choices are not liable to be sound. Internet of Things (IoT) quality rates big data with error, irregularity, deficiency, trickery, and model guess. Improving data veracity is critical to address these difficulties. In this article, we condense the key qualities and difficulties of IoT, which impact data handling and decision making. We audit the scene of estimating and upgrading data veracity and mining indeterminate data streams. Also, we propose five suggestions for future advancement of veracious big IoT data investigation that are identified with the heterogeneous and appropriated nature of IoT data, self-governing basic leadership, setting mindful and area streamlined philosophies, data cleaning and handling procedures for IoT edge gadgets, and protection safeguarding, customized, and secure data administration.  


Author(s):  
Hongzhi Wang ◽  
Jianzhong Li ◽  
Ran Huo ◽  
Li Jia ◽  
Lian Jin ◽  
...  

Author(s):  
Priya Deshpande ◽  
Alexander Rasin ◽  
Roselyne Tchoua ◽  
Jacob Furst ◽  
Daniela Raicu ◽  
...  
Keyword(s):  
Big Data ◽  

2016 ◽  
Vol 9 (3) ◽  
pp. 137-150 ◽  
Author(s):  
Zhao-Yang Qu ◽  
Yong-Wen Wang ◽  
Chong Wang ◽  
Nan Qu ◽  
Jia Yan

2018 ◽  
Vol 9 (2) ◽  
pp. 69-79 ◽  
Author(s):  
Klemen Kenda ◽  
Dunja Mladenić

Abstract Background: Internet of Things (IoT), earth observation and big scientific experiments are sources of extensive amounts of sensor big data today. We are faced with large amounts of data with low measurement costs. A standard approach in such cases is a stream mining approach, implying that we look at a particular measurement only once during the real-time processing. This requires the methods to be completely autonomous. In the past, very little attention was given to the most time-consuming part of the data mining process, i.e. data pre-processing. Objectives: In this paper we propose an algorithm for data cleaning, which can be applied to real-world streaming big data. Methods/Approach: We use the short-term prediction method based on the Kalman filter to detect admissible intervals for future measurements. The model can be adapted to the concept drift and is useful for detecting random additive outliers in a sensor data stream. Results: For datasets with low noise, our method has proven to perform better than the method currently commonly used in batch processing scenarios. Our results on higher noise datasets are comparable. Conclusions: We have demonstrated a successful application of the proposed method in real-world scenarios including the groundwater level, server load and smart-grid data


Sign in / Sign up

Export Citation Format

Share Document