scholarly journals Using naturalistic driving data to identify driving style based on longitudinal driving operation conditions

2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Nengchao Lyu ◽  
Yugang Wang ◽  
Chaozhong Wu ◽  
Lingfeng Peng ◽  
Alieu Freddie Thomas

Purpose An individual’s driving style significantly affects overall traffic safety. However, driving style is difficult to identify due to temporal and spatial differences and scene heterogeneity of driving behavior data. As such, the study of real-time driving-style identification methods is of great significance for formulating personalized driving strategies, improving traffic safety and reducing fuel consumption. This study aims to establish a driving style recognition framework based on longitudinal driving operation conditions (DOCs) using a machine learning model and natural driving data collected by a vehicle equipped with an advanced driving assistance system (ADAS). Design/methodology/approach Specifically, a driving style recognition framework based on longitudinal DOCs was established. To train the model, a real-world driving experiment was conducted. First, the driving styles of 44 drivers were preliminarily identified through natural driving data and video data; drivers were categorized through a subjective evaluation as conservative, moderate or aggressive. Then, based on the ADAS driving data, a criterion for extracting longitudinal DOCs was developed. Third, taking the ADAS data from 47 Kms of the two test expressways as the research object, six DOCs were calibrated and the characteristic data sets of the different DOCs were extracted and constructed. Finally, four machine learning classification (MLC) models were used to classify and predict driving style based on the natural driving data. Findings The results showed that six longitudinal DOCs were calibrated according to the proposed calibration criterion. Cautious drivers undertook the largest proportion of the free cruise condition (FCC), while aggressive drivers primarily undertook the FCC, following steady condition and relative approximation condition. Compared with cautious and moderate drivers, aggressive drivers adopted a smaller time headway (THW) and distance headway (DHW). THW, time-to-collision (TTC) and DHW showed highly significant differences in driving style identification, while longitudinal acceleration (LA) showed no significant difference in driving style identification. Speed and TTC showed no significant difference between moderate and aggressive drivers. In consideration of the cross-validation results and model prediction results, the overall hierarchical prediction performance ranking of the four studied machine learning models under the current sample data set was extreme gradient boosting > multi-layer perceptron > logistic regression > support vector machine. Originality/value The contribution of this research is to propose a criterion and solution for using longitudinal driving behavior data to label longitudinal DOCs and rapidly identify driving styles based on those DOCs and MLC models. This study provides a reference for real-time online driving style identification in vehicles equipped with onboard data acquisition equipment, such as ADAS.

2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Emmanuel Adinyira ◽  
Emmanuel Akoi-Gyebi Adjei ◽  
Kofi Agyekum ◽  
Frank Desmond Kofi Fugar

PurposeKnowledge of the effect of various cash-flow factors on expected project profit is important to effectively manage productivity on construction projects. This study was conducted to develop and test the sensitivity of a Machine Learning Support Vector Regression Algorithm (SVRA) to predict construction project profit in Ghana.Design/methodology/approachThe study relied on data from 150 institutional projects executed within the past five years (2014–2018) in developing the model. Eighty percent (80%) of the data from the 150 projects was used at hyperparameter selection and final training phases of the model development and the remaining 20% for model testing. Using MATLAB for Support Vector Regression, the parameters available for tuning were the epsilon values, the kernel scale, the box constraint and standardisations. The sensitivity index was computed to determine the degree to which the independent variables impact the dependent variable.FindingsThe developed model's predictions perfectly fitted the data and explained all the variability of the response data around its mean. Average predictive accuracy of 73.66% was achieved with all the variables on the different projects in validation. The developed SVR model was sensitive to labour and loan.Originality/valueThe developed SVRA combines variation, defective works and labour with other financial constraints, which have been the variables used in previous studies. It will aid contractors in predicting profit on completion at commencement and also provide information on the effect of changes to cash-flow factors on profit.


2022 ◽  
pp. ASN.2021040538
Author(s):  
Arthur M. Lee ◽  
Jian Hu ◽  
Yunwen Xu ◽  
Alison G. Abraham ◽  
Rui Xiao ◽  
...  

BackgroundUntargeted plasma metabolomic profiling combined with machine learning (ML) may lead to discovery of metabolic profiles that inform our understanding of pediatric CKD causes. We sought to identify metabolomic signatures in pediatric CKD based on diagnosis: FSGS, obstructive uropathy (OU), aplasia/dysplasia/hypoplasia (A/D/H), and reflux nephropathy (RN).MethodsUntargeted metabolomic quantification (GC-MS/LC-MS, Metabolon) was performed on plasma from 702 Chronic Kidney Disease in Children study participants (n: FSGS=63, OU=122, A/D/H=109, and RN=86). Lasso regression was used for feature selection, adjusting for clinical covariates. Four methods were then applied to stratify significance: logistic regression, support vector machine, random forest, and extreme gradient boosting. ML training was performed on 80% total cohort subsets and validated on 20% holdout subsets. Important features were selected based on being significant in at least two of the four modeling approaches. We additionally performed pathway enrichment analysis to identify metabolic subpathways associated with CKD cause.ResultsML models were evaluated on holdout subsets with receiver-operator and precision-recall area-under-the-curve, F1 score, and Matthews correlation coefficient. ML models outperformed no-skill prediction. Metabolomic profiles were identified based on cause. FSGS was associated with the sphingomyelin-ceramide axis. FSGS was also associated with individual plasmalogen metabolites and the subpathway. OU was associated with gut microbiome–derived histidine metabolites.ConclusionML models identified metabolomic signatures based on CKD cause. Using ML techniques in conjunction with traditional biostatistics, we demonstrated that sphingomyelin-ceramide and plasmalogen dysmetabolism are associated with FSGS and that gut microbiome–derived histidine metabolites are associated with OU.


2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Lei Li ◽  
Desheng Wu

PurposeThe infraction of securities regulations (ISRs) of listed firms in their day-to-day operations and management has become one of common problems. This paper proposed several machine learning approaches to forecast the risk at infractions of listed corporates to solve financial problems that are not effective and precise in supervision.Design/methodology/approachThe overall proposed research framework designed for forecasting the infractions (ISRs) include data collection and cleaning, feature engineering, data split, prediction approach application and model performance evaluation. We select Logistic Regression, Naïve Bayes, Random Forest, Support Vector Machines, Artificial Neural Network and Long Short-Term Memory Networks (LSTMs) as ISRs prediction models.FindingsThe research results show that prediction performance of proposed models with the prior infractions provides a significant improvement of the ISRs than those without prior, especially for large sample set. The results also indicate when judging whether a company has infractions, we should pay attention to novel artificial intelligence methods, previous infractions of the company, and large data sets.Originality/valueThe findings could be utilized to address the problems of identifying listed corporates' ISRs at hand to a certain degree. Overall, results elucidate the value of the prior infraction of securities regulations (ISRs). This shows the importance of including more data sources when constructing distress models and not only focus on building increasingly more complex models on the same data. This is also beneficial to the regulatory authorities.


2020 ◽  
Vol 7 (7) ◽  
pp. 2103
Author(s):  
Yoshihisa Matsunaga ◽  
Ryoichi Nakamura

Background: Abdominal cavity irrigation is a more minimally invasive surgery than that using a gas. Minimally invasive surgery improves the quality of life of patients; however, it demands higher skills from the doctors. Therefore, the study aimed to reduce the burden by assisting and automating the hemostatic procedure a highly frequent procedure by taking advantage of the clearness of the endoscopic images and continuous bleeding point observations in the liquid. We aimed to construct a method for detecting organs, bleeding sites, and hemostasis regions.Methods: We developed a method to perform real-time detection based on machine learning using laparoscopic videos. Our training dataset was prepared from three experiments in pigs. Linear support vector machine was applied using new color feature descriptors. In the verification of the accuracy of the classifier, we performed five-part cross-validation. Classification processing time was measured to verify the real-time property. Furthermore, we visualized the time series class change of the surgical field during the hemostatic procedure.Results: The accuracy of our classifier was 98.3% and the processing cost to perform real-time was enough. Furthermore, it was conceivable to quantitatively indicate the completion of the hemostatic procedure based on the changes in the bleeding region by ablation and the hemostasis regions by tissue coagulation.Conclusions: The organs, bleeding sites, and hemostasis regions classification was useful for assisting and automating the hemostatic procedure in the liquid. Our method can be adapted to more hemostatic procedures. 


Author(s):  
Zhi Zhang ◽  
Dagang Wang ◽  
Jianxiu Qiu ◽  
Jinxin Zhu ◽  
Tingli Wang

AbstractThe Global Precipitation Measurement (GPM) mission provides satellite precipitation products with an unprecedented spatio-temporal resolution and spatial coverage. However, its near-real-time (NRT) product still suffers from low accuracy. This study aims to improve the early run of the Integrated Multi-satellitE Retrievals for GPM (IMERG) by using four machine learning approaches, i.e., support vector machine (SVM), random forest (RF), artificial neural network (ANN), and Extreme Gradient Boosting (XGB). The cloud properties are selected as the predictors in addition to the original IMERG in these approaches. All the four approaches show similar improvement, with 53%-60% reduction of root-mean-square error (RMSE) compared with the original IMERG in a humid area, i.e., the Dongjiang River Basin (DJR) in southeastern China. The improvements are even greater in a semi-arid area, i.e., the Fenhe River Basin (FHR) in central China, the RMSE reduction ranges from 63%-66%. The products generated by the machine learning methods performs similarly to or even outperform than the final run of IMERG. Feature importance analysis, a technique to evaluate input features based on how useful they are in predicting a target variable, indicates that the cloud height and the brightness temperature are the most useful information in improving satellite precipitation products, followed by the atmospheric reflectivity and the surface temperature. This study shows that a more accurate NRT precipitation product can be produced by combining machine learning approaches and cloud information, which is of importance for hydrological applications that requires NRT precipitation information including flood monitoring.


2021 ◽  
Author(s):  
S. H. Al Gharbi ◽  
A. A. Al-Majed ◽  
A. Abdulraheem ◽  
S. Patil ◽  
S. M. Elkatatny

Abstract Due to high demand for energy, oil and gas companies started to drill wells in remote areas and unconventional environments. This raised the complexity of drilling operations, which were already challenging and complex. To adapt, drilling companies expanded their use of the real-time operation center (RTOC) concept, in which real-time drilling data are transmitted from remote sites to companies’ headquarters. In RTOC, groups of subject matter experts monitor the drilling live and provide real-time advice to improve operations. With the increase of drilling operations, processing the volume of generated data is beyond a human's capability, limiting the RTOC impact on certain components of drilling operations. To overcome this limitation, artificial intelligence and machine learning (AI/ML) technologies were introduced to monitor and analyze the real-time drilling data, discover hidden patterns, and provide fast decision-support responses. AI/ML technologies are data-driven technologies, and their quality relies on the quality of the input data: if the quality of the input data is good, the generated output will be good; if not, the generated output will be bad. Unfortunately, due to the harsh environments of drilling sites and the transmission setups, not all of the drilling data is good, which negatively affects the AI/ML results. The objective of this paper is to utilize AI/ML technologies to improve the quality of real-time drilling data. The paper fed a large real-time drilling dataset, consisting of over 150,000 raw data points, into Artificial Neural Network (ANN), Support Vector Machine (SVM) and Decision Tree (DT) models. The models were trained on the valid and not-valid datapoints. The confusion matrix was used to evaluate the different AI/ML models including different internal architectures. Despite the slowness of ANN, it achieved the best result with an accuracy of 78%, compared to 73% and 41% for DT and SVM, respectively. The paper concludes by presenting a process for using AI technology to improve real-time drilling data quality. To the author's knowledge based on literature in the public domain, this paper is one of the first to compare the use of multiple AI/ML techniques for quality improvement of real-time drilling data. The paper provides a guide for improving the quality of real-time drilling data.


2020 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Samira Khodabandehlou ◽  
S. Alireza Hashemi Golpayegani ◽  
Mahmoud Zivari Rahman

PurposeImproving the performance of recommender systems (RSs) has always been a major challenge in the area of e-commerce because the systems face issues such as cold start, sparsity, scalability and interest drift that affect their performance. Despite the efforts made to solve these problems, there is still no RS that can solve or reduce all the problems simultaneously. Therefore, the purpose of this study is to provide an effective and comprehensive RS to solve or reduce all of the above issues, which uses a combination of basic customer information as well as big data techniques.Design/methodology/approachThe most important steps in the proposed RS are: (1) collecting demographic and behavioral data of customers from an e-clothing store; (2) assessing customer personality traits; (3) creating a new user-item matrix based on customer/user interest; (4) calculating the similarity between customers with efficient k-nearest neighbor (EKNN) algorithm based on locality-sensitive hashing (LSH) approach and (5) defining a new similarity function based on a combination of personality traits, demographic characteristics and time-based purchasing behavior that are the key incentives for customers' purchases.FindingsThe proposed method was compared with different baselines (matrix factorization and ensemble). The results showed that the proposed method in terms of all evaluation measures led to a significant improvement in traditional collaborative filtering (CF) performance, and with a significant difference (more than 40%), performed better than all baselines. According to the results, we find that our proposed method, which uses a combination of personality information and demographics, as well as tracking the recent interests and needs of the customer with the LSH approach, helps to improve the effectiveness of the recommendations more than the baselines. This is due to the fact that this method, which uses the above information in conjunction with the LSH technique, is more effective and more accurate in solving problems of cold start, scalability, sparsity and interest drift.Research limitations/implicationsThe research data were limited to only one e-clothing store.Practical implicationsIn order to achieve an accurate and real-time RS in e-commerce, it is essential to use a combination of customer information with efficient techniques. In this regard, according to the results of the research, the use of personality traits and demographic characteristics lead to a more accurate knowledge of customers' interests and thus better identification of similar customers. Therefore, this information should be considered as a solution to reduce the problems of cold start and sparsity. Also, a better judgment can be made about customers' interests by considering their recent purchases; therefore, in order to solve the problems of interest drifts, different weights should be assigned to purchases and launch time of products/items at different times (the more recent, the more weight). Finally, the LSH technique is used to increase the RS scalability in e-commerce. In total, a combination of personality traits, demographics and customer purchasing behavior over time with the LSH technique should be used to achieve an ideal RS. Using the RS proposed in this research, it is possible to create a comfortable and enjoyable shopping experience for customers by providing real-time recommendations that match customers' preferences and can result in an increase in the profitability of e-shops.Originality/valueIn this study, by considering a combination of personality traits, demographic characteristics and time-based purchasing behavior of customers along with the LSH technique, we were able for the first time to simultaneously solve the basic problems of CF, namely cold start, scalability, sparsity and interest drift, which led to a decrease in significant errors of recommendations and an increase in the accuracy of CF. The average errors of the recommendations provided to users based on the proposed model is only about 13%, and the accuracy and compliance of these recommendations with the interests of customers is about 92%. In addition, a 40% difference between the accuracy of the proposed method and the traditional CF method has been observed. This level of accuracy in RSs is very significant and special, which is certainly welcomed by e-business owners. This is also a new scientific finding that is very useful for programmers, users and researchers. In general, the main contributions of this research are: 1) proposing an accurate RS using personality traits, demographic characteristics and time-based purchasing behavior; 2) proposing an effective and comprehensive RS for a “clothing” online store; 3) improving the RS performance by solving the cold start issue using personality traits and demographic characteristics; 4) improving the scalability issue in RS through efficient k-nearest neighbors; 5) Mitigating the sparsity issue by using personality traits and demographic characteristics and also by densifying the user-item matrix and 6) improving the RS accuracy by solving the interest drift issue through developing a time-based user-item matrix.


2020 ◽  
Vol 120 (6) ◽  
pp. 1149-1174 ◽  
Author(s):  
K.H. Leung ◽  
Daniel Y. Mo ◽  
G.T.S. Ho ◽  
C.H. Wu ◽  
G.Q. Huang

PurposeAccurate prediction of order demand across omni-channel supply chains improves the management's decision-making ability at strategic, tactical and operational levels. The paper aims to develop a predictive methodology for forecasting near-real-time e-commerce order arrivals in distribution centres, allowing third-party logistics service providers to manage the hour-to-hour fast-changing arrival rates of e-commerce orders better.Design/methodology/approachThe paper proposes a novel machine learning predictive methodology through the integration of the time series data characteristics into the development of an adaptive neuro-fuzzy inference system. A four-stage implementation framework is developed for enabling practitioners to apply the proposed model.FindingsA structured model evaluation framework is constructed for cross-validation of model performance. With the aid of an illustrative case study, forecasting evaluation reveals a high level of accuracy of the proposed machine learning approach in forecasting the arrivals of real e-commerce orders in three different retailers at three-hour intervals.Research limitations/implicationsResults from the case study suggest that real-time prediction of individual retailer's e-order arrival is crucial in order to maximize the value of e-order arrival prediction for daily operational decision-making.Originality/valueEarlier researchers examined supply chain demand, forecasting problem in a broader scope, particularly in dealing with the bullwhip effect. Prediction of real-time, hourly based order arrivals has been lacking. The paper fills this research gap by presenting a novel data-driven predictive methodology.


2017 ◽  
Vol 10 (2) ◽  
pp. 130-144 ◽  
Author(s):  
Iwan Aang Soenandi ◽  
Taufik Djatna ◽  
Ani Suryani ◽  
Irzaman Irzaman

Purpose The production of glycerol derivatives by the esterification process is subject to many constraints related to the yield of the production target and the lack of process efficiency. An accurate monitoring and controlling of the process can improve production yield and efficiency. The purpose of this paper is to propose a real-time optimization (RTO) using gradient adaptive selection and classification from infrared sensor measurement to cover various disturbances and uncertainties in the reactor. Design/methodology/approach The integration of the esterification process optimization using self-optimization (SO) was developed with classification process was combined with necessary condition optimum (NCO) as gradient adaptive selection, supported with laboratory scaled medium wavelength infrared (mid-IR) sensors, and measured the proposed optimization system indicator in the batch process. Business Process Modeling and Notation (BPMN 2.0) was built to describe the tasks of SO workflow in collaboration with NCO as an abstraction for the conceptual phase. Next, Stateflow modeling was deployed to simulate the three states of gradient-based adaptive control combined with support vector machine (SVM) classification and Arduino microcontroller for implementation. Findings This new method shows that the real-time optimization responsiveness of control increased product yield up to 13 percent, lower error measurement with percentage error 1.11 percent, reduced the process duration up to 22 minutes, with an effective range of stirrer rotation set between 300 and 400 rpm and final temperature between 200 and 210°C which was more efficient, as it consumed less energy. Research limitations/implications In this research the authors just have an experiment for the esterification process using glycerol, but as a development concept of RTO, it would be possible to apply for another chemical reaction or system. Practical implications This research introduces new development of an RTO approach to optimal control and as such marks the starting point for more research of its properties. As the methodology is generic, it can be applied to different optimization problems for a batch system in chemical industries. Originality/value The paper presented is original as it presents the first application of adaptive selection based on the gradient value of mid-IR sensor data, applied to the real-time determining control state by classification with the SVM algorithm for esterification process control to increase the efficiency.


Sensors ◽  
2021 ◽  
Vol 21 (19) ◽  
pp. 6349
Author(s):  
Jawad Ahmad ◽  
Johan Sidén ◽  
Henrik Andersson

This paper presents a posture recognition system aimed at detecting sitting postures of a wheelchair user. The main goals of the proposed system are to identify and inform irregular and improper posture to prevent sitting-related health issues such as pressure ulcers, with the potential that it could also be used for individuals without mobility issues. In the proposed monitoring system, an array of 16 screen printed pressure sensor units was employed to obtain pressure data, which are sampled and processed in real-time using read-out electronics. The posture recognition was performed for four sitting positions: right-, left-, forward- and backward leaning based on k-nearest neighbors (k-NN), support vector machines (SVM), random forest (RF), decision tree (DT) and LightGBM machine learning algorithms. As a result, a posture classification accuracy of up to 99.03 percent can be achieved. Experimental studies illustrate that the system can provide real-time pressure distribution value in the form of a pressure map on a standard PC and also on a raspberry pi system equipped with a touchscreen monitor. The stored pressure distribution data can later be shared with healthcare professionals so that abnormalities in sitting patterns can be identified by employing a post-processing unit. The proposed system could be used for risk assessments related to pressure ulcers. It may be served as a benchmark by recording and identifying individuals’ sitting patterns and the possibility of being realized as a lightweight portable health monitoring device.


Sign in / Sign up

Export Citation Format

Share Document