scholarly journals Boruta-grid-search least square support vector machine for NO2 pollution prediction using big data analytics and IoT emission sensors

2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Habeeb Balogun ◽  
Hafiz Alaka ◽  
Christian Nnaemeka Egwim

PurposeThis paper seeks to assess the performance levels of BA-GS-LSSVM compared to popular standalone algorithms used to build NO2 prediction models. The purpose of this paper is to pre-process a relatively large data of NO2 from Internet of Thing (IoT) sensors with time-corresponding weather and traffic data and to use the data to develop NO2 prediction models using BA-GS-LSSVM and popular standalone algorithms to allow for a fair comparison.Design/methodology/approachThis research installed and used data from 14 IoT emission sensors to develop machine learning predictive models for NO2 pollution concentration. The authors used big data analytics infrastructure to retrieve the large volume of data collected in tens of seconds for over 5 months. Weather data from the UK meteorology department and traffic data from the department for transport were collected and merged for the corresponding time and location where the pollution sensors exist.FindingsThe results show that the hybrid BA-GS-LSSVM outperforms all other standalone machine learning predictive Model for NO2 pollution.Practical implicationsThis paper's hybrid model provides a basis for giving an informed decision on the NO2 pollutant avoidance system.Originality/valueThis research installed and used data from 14 IoT emission sensors to develop machine learning predictive models for NO2 pollution concentration.

2022 ◽  
Vol 2022 ◽  
pp. 1-10
Author(s):  
Raja Krishnamoorthi ◽  
Shubham Joshi ◽  
Hatim Z. Almarzouki ◽  
Piyush Kumar Shukla ◽  
Ali Rizwan ◽  
...  

Diabetes is a chronic disease that continues to be a significant and global concern since it affects the entire population’s health. It is a metabolic disorder that leads to high blood sugar levels and many other problems such as stroke, kidney failure, and heart and nerve problems. Several researchers have attempted to construct an accurate diabetes prediction model over the years. However, this subject still faces significant open research issues due to a lack of appropriate data sets and prediction approaches, which pushes researchers to use big data analytics and machine learning (ML)-based methods. Applying four different machine learning methods, the research tries to overcome the problems and investigate healthcare predictive analytics. The study’s primary goal was to see how big data analytics and machine learning-based techniques may be used in diabetes. The examination of the results shows that the suggested ML-based framework may achieve a score of 86. Health experts and other stakeholders are working to develop categorization models that will aid in the prediction of diabetes and the formulation of preventative initiatives. The authors perform a review of the literature on machine models and suggest an intelligent framework for diabetes prediction based on their findings. Machine learning models are critically examined, and an intelligent machine learning-based architecture for diabetes prediction is proposed and evaluated by the authors. In this study, the authors utilize our framework to develop and assess decision tree (DT)-based random forest (RF) and support vector machine (SVM) learning models for diabetes prediction, which are the most widely used techniques in the literature at the time of writing. It is proposed in this study that a unique intelligent diabetes mellitus prediction framework (IDMPF) is developed using machine learning. According to the framework, it was developed after conducting a rigorous review of existing prediction models in the literature and examining their applicability to diabetes. Using the framework, the authors describe the training procedures, model assessment strategies, and issues associated with diabetes prediction, as well as solutions they provide. The findings of this study may be utilized by health professionals, stakeholders, students, and researchers who are involved in diabetes prediction research and development. The proposed work gives 83% accuracy with the minimum error rate.


Webology ◽  
2021 ◽  
Vol 18 (Special Issue 04) ◽  
pp. 591-606
Author(s):  
R. Brindha ◽  
Dr.M. Thillaikarasi

Big data analytics (BDA) is a system based method with an aim to recognize and examine different designs, patterns and trends under the big dataset. In this paper, BDA is used to visualize and trends the prediction where exploratory data analysis examines the crime data. “A successive facts and patterns have been taken in following cities of California, Washington and Florida by using statistical analysis and visualization”. The predictive result gives the performance using Keras Prophet Model, LSTM and neural network models followed by prophet model which are the existing methods used to find the crime data under BDA technique. But the crime actions increases day by day which is greater task for the people to overcome the challenging crime activities. Some ignored the essential rate of influential aspects. To overcome these challenging problems of big data, many studies have been developed with limited one or two features. “This paper introduces a big data introduces to analyze the influential aspects about the crime incidents, and examine it on New York City. The proposed structure relates the dynamic machine learning algorithms and geographical information system (GIS) to consider the contiguous reasons of crime data. Recursive feature elimination (RFE) is used to select the optimum characteristic data. Exploitation of gradient boost decision tree (GBDT), logistic regression (LR), support vector machine (SVM) and artificial neural network (ANN) are related to develop the optimum data model. Significant impact features were then reviewed by applying GBDT and GIS”. The experimental results illustrates that GBDT along with GIS model combination can identify the crime ranking with high performance and accuracy compared to existing method.”


2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Minwoo Lee ◽  
Wooseok Kwon ◽  
Ki-Joon Back

Purpose Big data analytics allows researchers and industry practitioners to extract hidden patterns or discover new information and knowledge from big data. Although artificial intelligence (AI) is one of the emerging big data analytics techniques, hospitality and tourism literature has shown minimal efforts to process and analyze big hospitality data through AI. Thus, this study aims to develop and compare prediction models for review helpfulness using machine learning (ML) algorithms to analyze big restaurant data. Design/methodology/approach The study analyzed 1,483,858 restaurant reviews collected from Yelp.com. After a thorough literature review, the study identified and added to the prediction model 4 attributes containing 11 key determinants of review helpfulness. Four ML algorithms, namely, multivariate linear regression, random forest, support vector machine regression and extreme gradient boosting (XGBoost), were used to find a better prediction model for customer decision-making. Findings By comparing the performance metrics, the current study found that XGBoost was the best model to predict review helpfulness among selected popular ML algorithms. Results revealed that attributes regarding a reviewer’s credibility were fundamental factors determining a review’s helpfulness. Review helpfulness even valued credibility over ratings or linguistic contents such as sentiment and subjectivity. Practical implications The current study helps restaurant operators to attract customers by predicting review helpfulness through ML-based predictive modeling and presenting potential helpful reviews based on critical attributes including review, reviewer, restaurant and linguistic content. Using AI, online review platforms and restaurant websites can enhance customers’ attitude and purchase decision-making by reducing information overload and search cost and highlighting the most crucial review helpfulness features and user-friendly automated search results. Originality/value To the best of the authors’ knowledge, the current study is the first to develop a prediction model of review helpfulness and reveal essential factors for helpful reviews. Furthermore, the study presents a state-of-the-art ML model that surpasses the conventional models’ prediction accuracy. The findings will improve practitioners’ marketing strategies by focusing on factors that influence customers’ decision-making.


2019 ◽  
Vol 32 (2) ◽  
pp. 82-92 ◽  
Author(s):  
Sung Yi ◽  
Robert Jones

Purpose This paper aims to present a machine learning framework for using big data analytics to predict the reliability of solder joints. The purpose of this study is to accurately predict the reliability of solder joints by using big data analytics. Design/methodology/approach A machine learning framework for using big data analytics is proposed to predict the reliability of solder joints accurately. Findings A machine learning framework for predicting the life of solder joints accurately has been developed in this study. To validate its accuracy and efficiency, it is applied to predict the long-term reliability of lead-free Sn96.5Ag3.0Cu0.5 (SAC305) for three commonly used surface finishes such OSP, ENIG and IAg. The obtained results show that the predicted failure based on the machine learning method is much more accurate than the Weibull method. In addition, solder ball/bump joint failure modes are identified based on various solder joint failures reported in the literature. Originality/value The ability to predict thermal fatigue life accurately is extremely valuable to the industry because it saves time and cost for product development and optimization.


2017 ◽  
Vol 21 (1) ◽  
pp. 12-17 ◽  
Author(s):  
David J. Pauleen

Purpose Dave Snowden has been an important voice in knowledge management over the years. As the founder and chief scientific officer of Cognitive Edge, a company focused on the development of the theory and practice of social complexity, he offers informative views on the relationship between big data/analytics and KM. Design/methodology/approach A face-to-face interview was held with Dave Snowden in May 2015 in Auckland, New Zealand. Findings According to Snowden, analytics in the form of algorithms are imperfect and can only to a small extent capture the reasoning and analytical capabilities of people. For this reason, while big data/analytics can be useful, they are limited and must be used in conjunction with human knowledge and reasoning. Practical implications Snowden offers his views on big data/analytics and how they can be used effectively in real world situations in combination with human reasoning and input, for example in fields from resource management to individual health care. Originality/value Snowden is an innovative thinker. He combines knowledge and experience from many fields and offers original views and understanding of big data/analytics, knowledge and management.


2020 ◽  
Vol 102 (913) ◽  
pp. 199-234
Author(s):  
Nema Milaninia

AbstractAdvances in mobile phone technology and social media have created a world where the volume of information generated and shared is outpacing the ability of humans to review and use that data. Machine learning (ML) models and “big data” analytical tools have the power to ease that burden by making sense of this information and providing insights that might not otherwise exist. In the context of international criminal and human rights law, ML is being used for a variety of purposes, including to uncover mass graves in Mexico, find evidence of homes and schools destroyed in Darfur, detect fake videos and doctored evidence, predict the outcomes of judicial hearings at the European Court of Human Rights, and gather evidence of war crimes in Syria. ML models are also increasingly being incorporated by States into weapon systems in order to better enable targeting systems to distinguish between civilians, allied soldiers and enemy combatants or even inform decision-making for military attacks.The same technology, however, also comes with significant risks. ML models and big data analytics are highly susceptible to common human biases. As a result of these biases, ML models have the potential to reinforce and even accelerate existing racial, political or gender inequalities, and can also paint a misleading and distorted picture of the facts on the ground. This article discusses how common human biases can impact ML models and big data analytics, and examines what legal implications these biases can have under international criminal law and international humanitarian law.


Sign in / Sign up

Export Citation Format

Share Document