scholarly journals Managing Uncertainty in Geological Scenarios Using Machine Learning-Based Classification Model on Production Data

Geofluids ◽  
2020 ◽  
Vol 2020 ◽  
pp. 1-16 ◽  
Author(s):  
Byeongcheol Kang ◽  
Kyungbook Lee

Training image (TI) has a great influence on reservoir modeling as a spatial correlation in the multipoint geostatistics. Unlike the variogram of the two-point geostatistics that is mathematically defined, there is a high degree of geological uncertainty to determine a proper TI. The goal of this study is to develop a classification model for determining the proper geological scenario among plausible TIs by using machine learning methods: (a) support vector machine (SVM), (b) artificial neural network (ANN), and (c) convolutional neural network (CNN). After simulated production data are used to train the classification model, the most possible TI can be selected when the observed production responses are put into the trained model. This study, as far as we know, is the first application of CNN in which production history data are composed as a matrix form for use as an input image. The training data are set to cover various production trends to make the machine learning models more reliable. Therefore, a total of 800 channelized reservoirs were generated from four TIs, which have different channel directions to consider geological uncertainty. We divided them into training, validation, and test sets of 576, 144, and 80, respectively. The input layer comprised 800 production data, i.e., oil production rates and water cuts for eight production wells over 50 time steps, and the output layer consisted of a probability vector for each TI. The SVM and CNN models reasonably reduced the uncertainty in modeling the facies distribution based on the reliable probability for each TI. Even though the ANN and CNN had roughly the same number of parameters, the CNN outperformed the ANN in terms of both validation and test sets. The CNN successfully classified the reference model’s TI with about 95% probability. This is because the CNN can grasp the overall trend of production history. The probabilities of TI from the SVM and CNN were applied to regenerate more reliable reservoir models using the concept of TI rejection and reduced the uncertainty in the geological scenario successfully.

Sensors ◽  
2020 ◽  
Vol 20 (24) ◽  
pp. 7212
Author(s):  
Jungryul Seo ◽  
Teemu H. Laine ◽  
Gyuhwan Oh ◽  
Kyung-Ah Sohn

As the number of patients with Alzheimer’s disease (AD) increases, the effort needed to care for these patients increases as well. At the same time, advances in information and sensor technologies have reduced caring costs, providing a potential pathway for developing healthcare services for AD patients. For instance, if a virtual reality (VR) system can provide emotion-adaptive content, the time that AD patients spend interacting with VR content is expected to be extended, allowing caregivers to focus on other tasks. As the first step towards this goal, in this study, we develop a classification model that detects AD patients’ emotions (e.g., happy, peaceful, or bored). We first collected electroencephalography (EEG) data from 30 Korean female AD patients who watched emotion-evoking videos at a medical rehabilitation center. We applied conventional machine learning algorithms, such as a multilayer perceptron (MLP) and support vector machine, along with deep learning models of recurrent neural network (RNN) architectures. The best performance was obtained from MLP, which achieved an average accuracy of 70.97%; the RNN model’s accuracy reached only 48.18%. Our study results open a new stream of research in the field of EEG-based emotion detection for patients with neurological disorders.


2020 ◽  
Vol 13 (1-2) ◽  
pp. 43-52
Author(s):  
Boudewijn van Leeuwen ◽  
Zalán Tobak ◽  
Ferenc Kovács

AbstractClassification of multispectral optical satellite data using machine learning techniques to derive land use/land cover thematic data is important for many applications. Comparing the latest algorithms, our research aims to determine the best option to classify land use/land cover with special focus on temporary inundated land in a flat area in the south of Hungary. These inundations disrupt agricultural practices and can cause large financial loss. Sentinel 2 data with a high temporal and medium spatial resolution is classified using open source implementations of a random forest, support vector machine and an artificial neural network. Each classification model is applied to the same data set and the results are compared qualitatively and quantitatively. The accuracy of the results is high for all methods and does not show large overall differences. A quantitative spatial comparison demonstrates that the neural network gives the best results, but that all models are strongly influenced by atmospheric disturbances in the image.


2021 ◽  
Vol 11 (22) ◽  
pp. 10682
Author(s):  
Pham-The Hien ◽  
Ic-Pyo Hong

Wall-thinning in building structures due to corrosion and surface erosion occurs due to the severe operating conditions and the changing of the surrounding environment, or it can result from poor workmanship and a lack of systematic monitoring during construction. Hence, the continuous monitoring of structures plays an important role in decreasing unexpected accidents. In this paper, a novel method based on the deep neural network and support vector machine approaches is investigated to build up a thickness classification model by incorporating different input features, including the dielectric constants of the material under test, which are extracted from the scattering parameters proceeded by the National Institute of Standards and Technology iterative method. The attained classification results from both machine learning algorithms are then compared and show that both of the models have a good prediction ability. While the deep neural network is the better solution with a large amount of data, the support vector machine is the more appropriate solution when employing small dataset. It can be stated that the proposed method is able to support systematic monitoring as it can help to improve the accuracy of the prediction of material thickness.


2021 ◽  
Vol 6 (2) ◽  
Author(s):  
Solomon Akinboro ◽  
Isaac K. Ogundoyin ◽  
Ayobami T. Olusesi

Machine learning has been an effective tool to connect networks of enormous information for predicting personality. Identification of personality-related indicators encrypted in Facebook profiles and activities are of special concern in most research efforts. This research modeled user personality based on set of features extracted from the Facebook data using Map-Reduce Back Propagation Neural Network (MRBPNN). The performance of the MRBPNN classification model was evaluated in terms of five basic personality dimensions: Extraversion (EXT), Agreeableness (AGR), Conscientiousness (CON), Neuroticism (NEU), and Openness to Experience (OPN) using True positive, False Positive, accuracy, precision and F-measure as metrics at the threshold value of 0.32. The experimental results reveal that MRBPNN model has accuracy of 91.40%, 93.89%, 91.33%, 90.43% and 89.13% CON, OPN, EXT, NEU and AGR respectively for personality recognition which is more computationally efficient than Back Propagation Neural Network (BPNN) and Support Vector Machine (SVM). Therefore, personality recognition based on MRBPNN would produce a reliable prediction system for various personality traits with data having a very large instance.  Keywords— Machine learning, Facebook, MRBPNN, Personality Recognition, Neuroticism, Agreeableness.


Electronics ◽  
2021 ◽  
Vol 10 (14) ◽  
pp. 1694
Author(s):  
Mathew Ashik ◽  
A. Jyothish ◽  
S. Anandaram ◽  
P. Vinod ◽  
Francesco Mercaldo ◽  
...  

Malware is one of the most significant threats in today’s computing world since the number of websites distributing malware is increasing at a rapid rate. Malware analysis and prevention methods are increasingly becoming necessary for computer systems connected to the Internet. This software exploits the system’s vulnerabilities to steal valuable information without the user’s knowledge, and stealthily send it to remote servers controlled by attackers. Traditionally, anti-malware products use signatures for detecting known malware. However, the signature-based method does not scale in detecting obfuscated and packed malware. Considering that the cause of a problem is often best understood by studying the structural aspects of a program like the mnemonics, instruction opcode, API Call, etc. In this paper, we investigate the relevance of the features of unpacked malicious and benign executables like mnemonics, instruction opcodes, and API to identify a feature that classifies the executable. Prominent features are extracted using Minimum Redundancy and Maximum Relevance (mRMR) and Analysis of Variance (ANOVA). Experiments were conducted on four datasets using machine learning and deep learning approaches such as Support Vector Machine (SVM), Naïve Bayes, J48, Random Forest (RF), and XGBoost. In addition, we also evaluate the performance of the collection of deep neural networks like Deep Dense network, One-Dimensional Convolutional Neural Network (1D-CNN), and CNN-LSTM in classifying unknown samples, and we observed promising results using APIs and system calls. On combining APIs/system calls with static features, a marginal performance improvement was attained comparing models trained only on dynamic features. Moreover, to improve accuracy, we implemented our solution using distinct deep learning methods and demonstrated a fine-tuned deep neural network that resulted in an F1-score of 99.1% and 98.48% on Dataset-2 and Dataset-3, respectively.


Sensors ◽  
2021 ◽  
Vol 21 (9) ◽  
pp. 3068
Author(s):  
Soumaya Dghim ◽  
Carlos M. Travieso-González ◽  
Radim Burget

The use of image processing tools, machine learning, and deep learning approaches has become very useful and robust in recent years. This paper introduces the detection of the Nosema disease, which is considered to be one of the most economically significant diseases today. This work shows a solution for recognizing and identifying Nosema cells between the other existing objects in the microscopic image. Two main strategies are examined. The first strategy uses image processing tools to extract the most valuable information and features from the dataset of microscopic images. Then, machine learning methods are applied, such as a neural network (ANN) and support vector machine (SVM) for detecting and classifying the Nosema disease cells. The second strategy explores deep learning and transfers learning. Several approaches were examined, including a convolutional neural network (CNN) classifier and several methods of transfer learning (AlexNet, VGG-16 and VGG-19), which were fine-tuned and applied to the object sub-images in order to identify the Nosema images from the other object images. The best accuracy was reached by the VGG-16 pre-trained neural network with 96.25%.


Animals ◽  
2020 ◽  
Vol 10 (5) ◽  
pp. 771
Author(s):  
Toshiya Arakawa

Mammalian behavior is typically monitored by observation. However, direct observation requires a substantial amount of effort and time, if the number of mammals to be observed is sufficiently large or if the observation is conducted for a prolonged period. In this study, machine learning methods as hidden Markov models (HMMs), random forests, support vector machines (SVMs), and neural networks, were applied to detect and estimate whether a goat is in estrus based on the goat’s behavior; thus, the adequacy of the method was verified. Goat’s tracking data was obtained using a video tracking system and used to estimate whether they, which are in “estrus” or “non-estrus”, were in either states: “approaching the male”, or “standing near the male”. Totally, the PC of random forest seems to be the highest. However, The percentage concordance (PC) value besides the goats whose data were used for training data sets is relatively low. It is suggested that random forest tend to over-fit to training data. Besides random forest, the PC of HMMs and SVMs is high. However, considering the calculation time and HMM’s advantage in that it is a time series model, HMM is better method. The PC of neural network is totally low, however, if the more goat’s data were acquired, neural network would be an adequate method for estimation.


2021 ◽  
Author(s):  
Rui Liu ◽  
Xin Yang ◽  
Chong Xu ◽  
Luyao Li ◽  
Xiangqiang Zeng

Abstract Landslide susceptibility mapping (LSM) is a useful tool to estimate the probability of landslide occurrence, providing a scientific basis for natural hazards prevention, land use planning, and economic development in landslide-prone areas. To date, a large number of machine learning methods have been applied to LSM, and recently the advanced Convolutional Neural Network (CNN) has been gradually adopted to enhance the prediction accuracy of LSM. The objective of this study is to introduce a CNN based model in LSM and systematically compare its overall performance with the conventional machine learning models of random forest, logistic regression, and support vector machine. Herein, we selected the Jiuzhaigou region in Sichuan Province, China as the study area. A total number of 710 landslides and 12 predisposing factors were stacked to form spatial datasets for LSM. The ROC analysis and several statistical metrics, such as accuracy, root mean square error (RMSE), Kappa coefficient, sensitivity, and specificity were used to evaluate the performance of the models in the training and validation datasets. Finally, the trained models were calculated and the landslide susceptibility zones were mapped. Results suggest that both CNN and conventional machine-learning based models have a satisfactory performance (AUC: 85.72% − 90.17%). The CNN based model exhibits excellent good-of-fit and prediction capability, and achieves the highest performance (AUC: 90.17%) but also significantly reduces the salt-of-pepper effect, which indicates its great potential of application to LSM.


2020 ◽  
Vol 10 (18) ◽  
pp. 6417 ◽  
Author(s):  
Emanuele Lattanzi ◽  
Giacomo Castellucci ◽  
Valerio Freschi

Most road accidents occur due to human fatigue, inattention, or drowsiness. Recently, machine learning technology has been successfully applied to identifying driving styles and recognizing unsafe behaviors starting from in-vehicle sensors signals such as vehicle and engine speed, throttle position, and engine load. In this work, we investigated the fusion of different external sensors, such as a gyroscope and a magnetometer, with in-vehicle sensors, to increase machine learning identification of unsafe driver behavior. Starting from those signals, we computed a set of features capable to accurately describe the behavior of the driver. A support vector machine and an artificial neural network were then trained and tested using several features calculated over more than 200 km of travel. The ground truth used to evaluate classification performances was obtained by means of an objective methodology based on the relationship between speed, and lateral and longitudinal acceleration of the vehicle. The classification results showed an average accuracy of about 88% using the SVM classifier and of about 90% using the neural network demonstrating the potential capability of the proposed methodology to identify unsafe driver behaviors.


Sensors ◽  
2021 ◽  
Vol 21 (23) ◽  
pp. 7853
Author(s):  
Aleksej Logacjov ◽  
Kerstin Bach ◽  
Atle Kongsvold ◽  
Hilde Bremseth Bårdstu ◽  
Paul Jarle Mork

Existing accelerometer-based human activity recognition (HAR) benchmark datasets that were recorded during free living suffer from non-fixed sensor placement, the usage of only one sensor, and unreliable annotations. We make two contributions in this work. First, we present the publicly available Human Activity Recognition Trondheim dataset (HARTH). Twenty-two participants were recorded for 90 to 120 min during their regular working hours using two three-axial accelerometers, attached to the thigh and lower back, and a chest-mounted camera. Experts annotated the data independently using the camera’s video signal and achieved high inter-rater agreement (Fleiss’ Kappa =0.96). They labeled twelve activities. The second contribution of this paper is the training of seven different baseline machine learning models for HAR on our dataset. We used a support vector machine, k-nearest neighbor, random forest, extreme gradient boost, convolutional neural network, bidirectional long short-term memory, and convolutional neural network with multi-resolution blocks. The support vector machine achieved the best results with an F1-score of 0.81 (standard deviation: ±0.18), recall of 0.85±0.13, and precision of 0.79±0.22 in a leave-one-subject-out cross-validation. Our highly professional recordings and annotations provide a promising benchmark dataset for researchers to develop innovative machine learning approaches for precise HAR in free living.


Sign in / Sign up

Export Citation Format

Share Document