Evaluating the Effectiveness of Machine Learning Technologies in Improving Real-Time Drilling Data-Quality

2022 ◽  
pp. 1-14
Author(s):  
Salem Al-Gharbi ◽  
Abdulaziz Al-Majed ◽  
Salaheldin Elkatatny ◽  
Abdulazeez Abdulraheem

Abstract Due to high demand for energy, oil and gas companies started to drill wells in remote environments conducting unconventional operations. In order to maintain safe, fast and more cost-effective operations, utilizing machine learning (ML) technologies has become a must. The harsh environments of drilling sites and the transmission setups, are negatively affecting the drilling data, leading to less than acceptable ML results. For that reason, big portion of ML development projects were actually spent on improving the data by data-quality experts. The objective of this paper is to evaluate the effectiveness of ML on improving the real-time drilling-data-quality and compare it to a human expert knowledge. To achieve that, two large real-time drilling datasets were used; one dataset was used to train three different ML techniques: artificial neural network (ANN), support vector machine (SVM) and decision tree (DT), the second dataset was used to evaluate it. The ML results were compared with the results of a real- time drilling data quality expert. Despite the complexity of ANN and good results in general, it achieved a relative root mean square error (RRMSE) of 2.83%, which was lower than DT and SVM technologies that achieved RRMSE of 0.35% and 0.48% respectively. The uniqueness of this work is in developing ML that simulates the improvement of drilling-data- quality by an expert. This research provides a guide for improving the quality of real-time drilling data.

2021 ◽  
Author(s):  
S. H. Al Gharbi ◽  
A. A. Al-Majed ◽  
A. Abdulraheem ◽  
S. Patil ◽  
S. M. Elkatatny

Abstract Due to high demand for energy, oil and gas companies started to drill wells in remote areas and unconventional environments. This raised the complexity of drilling operations, which were already challenging and complex. To adapt, drilling companies expanded their use of the real-time operation center (RTOC) concept, in which real-time drilling data are transmitted from remote sites to companies’ headquarters. In RTOC, groups of subject matter experts monitor the drilling live and provide real-time advice to improve operations. With the increase of drilling operations, processing the volume of generated data is beyond a human's capability, limiting the RTOC impact on certain components of drilling operations. To overcome this limitation, artificial intelligence and machine learning (AI/ML) technologies were introduced to monitor and analyze the real-time drilling data, discover hidden patterns, and provide fast decision-support responses. AI/ML technologies are data-driven technologies, and their quality relies on the quality of the input data: if the quality of the input data is good, the generated output will be good; if not, the generated output will be bad. Unfortunately, due to the harsh environments of drilling sites and the transmission setups, not all of the drilling data is good, which negatively affects the AI/ML results. The objective of this paper is to utilize AI/ML technologies to improve the quality of real-time drilling data. The paper fed a large real-time drilling dataset, consisting of over 150,000 raw data points, into Artificial Neural Network (ANN), Support Vector Machine (SVM) and Decision Tree (DT) models. The models were trained on the valid and not-valid datapoints. The confusion matrix was used to evaluate the different AI/ML models including different internal architectures. Despite the slowness of ANN, it achieved the best result with an accuracy of 78%, compared to 73% and 41% for DT and SVM, respectively. The paper concludes by presenting a process for using AI technology to improve real-time drilling data quality. To the author's knowledge based on literature in the public domain, this paper is one of the first to compare the use of multiple AI/ML techniques for quality improvement of real-time drilling data. The paper provides a guide for improving the quality of real-time drilling data.


2021 ◽  
Author(s):  
Temirlan Zhekenov ◽  
Artem Nechaev ◽  
Kamilla Chettykbayeva ◽  
Alexey Zinovyev ◽  
German Sardarov ◽  
...  

SUMMARY Researchers base their analysis on basic drilling parameters obtained during mud logging and demonstrate impressive results. However, due to limitations imposed by data quality often present during drilling, those solutions often tend to lose their stability and high levels of predictivity. In this work, the concept of hybrid modeling was introduced which allows to integrate the analytical correlations with algorithms of machine learning for obtaining stable solutions consistent from one data set to another.


2021 ◽  
Author(s):  
Vagif Suleymanov ◽  
Hany Gamal ◽  
Guenther Glatz ◽  
Salaheldin Elkatatny ◽  
Abdulazeez Abdulraheem

Abstract Acoustic data obtained from sonic logging tools plays an important role in formation evaluation. Given the associated costs, however, the industry clearly stands to benefit from cheaper technologies to obtain compressional and shear wave slowness data. Therefore, this paper delineates an alternative solution for the prediction of sonic log data by means of Machine Learning (ML). This study takes advantage of an adaptive neuro-fuzzy inference system (ANFIS) and support vector machine (SVM) ML techniques to predict compressional and shear wave slowness from drilling data only. In particular, the network is trained utilizing 2000 data points such as weight on bit (WOB), rate of penetration (ROP), standpipe pressure (SPP), torque (T), drill pipe rotation (RPM), and mud flow rate (GPM). Consequently, acoustic properties of the rock can be estimated solely from readily available parameters thereby saving both costs and time associated with sonic logs. The obtained results are promising and supportive of both ANFIS and SVM model as viable alternatives to obtain sonic data without the need for running sonic logs. The developed ANFIS model was able to predict compressional and shear wave slowness with correlation coefficients of 0.94 and 0.98 and average absolute percentage errors (AAPE) of 1.87% and 2.61%, respectively. Similarly, the SVM model predicted sonic logs with high accuracy yielding to correlation coefficients of more than 0.98 and AAPE of 0.74% and 0.84% for both compressional and shear logs, respectively. Once a network is trained, the approach naturally lends itself to be integrated as a real time service. This study outlines a novel and cost-effective solution to estimate rock compressional and shear-wave slowness solely from readily available drilling parameters. Importantly, the model has been verified for wells drilled in different formations with complex lithology substantiating the effectiveness of the approach.


2021 ◽  
Author(s):  
ADNAN ZAHID ◽  
Kia Dashtipour ◽  
Ivonne E. Carranza1 ◽  
Hasan Abbas ◽  
Aifeng Ren ◽  
...  

Abstract Water is considered to be the most essential and vital resources to sustaining life. Ensuring its delivery to people with no intrusion of harmful impurities, safe, reliable, and in an affordable manner is one of huge challenge amid to the ongoing climate transformations. This demands to introduce a cost effective and notion of real-time monitoring system that can detect the microbiological contaminants in aqueous solutions in timely manner to protect the public and environment health. In this paper, the prospects of integrating non-invasive terahertz (THz) waves with machine learning (ML) enabled technique is studied. The research explores a method of using Fourier transform Infrared Spectroscopy (FTIR) system to observe the absorption spectra and characteristics of three solvents solution , including salt, sugar and glucose with various quantity in aqueous solutions in the frequency range of 1 THz to 20 THz. In this study, due to the different molecular configuration and vibration modes of substances, distinct absorption spectra peaks were achieved for different concentrations of solvent solutions at certain sensitive THz region. Moreover, using measurements observations data, meaningful features are extracted and incorporated four algorithms such as random forest (RF), support vector machine (SVM), decision tree (D-tree) and k-nearest neighbour (KNN). The results demonstrated that RF obtained a higher accuracy of 84.74% in identifying the substance in aqueous solutions. Moreover, it was also found that RF with 97.98%, outperformed other classifiers for estimation of salts concentration added in aqueous solutions. However, for sugar and glucose concentrations, SVM exhibited a higher accuracy of 93.11% and 96.88%, respectively, compared to other classifiers. Thus, proposed technique incorporating ML with THz waves, may be significant in providing an efficient, cost-effective and real-time monitoring for water quality detection system.


2021 ◽  
Vol 7 ◽  
pp. e813
Author(s):  
Anandan Chinnalagu ◽  
Ashok Kumar Durairaj

Customer satisfaction and their positive sentiments are some of the various goals for successful companies. However, analyzing customer reviews to predict accurate sentiments have been proven to be challenging and time-consuming due to high volumes of collected data from various sources. Several researchers approach this with algorithms, methods, and models. These include machine learning and deep learning (DL) methods, unigram and skip-gram based algorithms, as well as the Artificial Neural Network (ANN) and bag-of-word (BOW) regression model. Studies and research have revealed incoherence in polarity, model overfitting and performance issues, as well as high cost in data processing. This experiment was conducted to solve these revealing issues, by building a high performance yet cost-effective model for predicting accurate sentiments from large datasets containing customer reviews. This model uses the fastText library from Facebook’s AI research (FAIR) Lab, as well as the traditional Linear Support Vector Machine (LSVM) to classify text and word embedding. Comparisons of this model were also done with the author’s a custom multi-layer Sentiment Analysis (SA) Bi-directional Long Short-Term Memory (SA-BLSTM) model. The proposed fastText model, based on results, obtains a higher accuracy of 90.71% as well as 20% in performance compared to LSVM and SA-BLSTM models.


2019 ◽  
Vol 20 (5) ◽  
pp. 488-500 ◽  
Author(s):  
Yan Hu ◽  
Yi Lu ◽  
Shuo Wang ◽  
Mengying Zhang ◽  
Xiaosheng Qu ◽  
...  

Background: Globally the number of cancer patients and deaths are continuing to increase yearly, and cancer has, therefore, become one of the world&#039;s highest causes of morbidity and mortality. In recent years, the study of anticancer drugs has become one of the most popular medical topics. </P><P> Objective: In this review, in order to study the application of machine learning in predicting anticancer drugs activity, some machine learning approaches such as Linear Discriminant Analysis (LDA), Principal components analysis (PCA), Support Vector Machine (SVM), Random forest (RF), k-Nearest Neighbor (kNN), and Naïve Bayes (NB) were selected, and the examples of their applications in anticancer drugs design are listed. </P><P> Results: Machine learning contributes a lot to anticancer drugs design and helps researchers by saving time and is cost effective. However, it can only be an assisting tool for drug design. </P><P> Conclusion: This paper introduces the application of machine learning approaches in anticancer drug design. Many examples of success in identification and prediction in the area of anticancer drugs activity prediction are discussed, and the anticancer drugs research is still in active progress. Moreover, the merits of some web servers related to anticancer drugs are mentioned.


2021 ◽  
Vol 186 (Supplement_1) ◽  
pp. 445-451
Author(s):  
Yifei Sun ◽  
Navid Rashedi ◽  
Vikrant Vaze ◽  
Parikshit Shah ◽  
Ryan Halter ◽  
...  

ABSTRACT Introduction Early prediction of the acute hypotensive episode (AHE) in critically ill patients has the potential to improve outcomes. In this study, we apply different machine learning algorithms to the MIMIC III Physionet dataset, containing more than 60,000 real-world intensive care unit records, to test commonly used machine learning technologies and compare their performances. Materials and Methods Five classification methods including K-nearest neighbor, logistic regression, support vector machine, random forest, and a deep learning method called long short-term memory are applied to predict an AHE 30 minutes in advance. An analysis comparing model performance when including versus excluding invasive features was conducted. To further study the pattern of the underlying mean arterial pressure (MAP), we apply a regression method to predict the continuous MAP values using linear regression over the next 60 minutes. Results Support vector machine yields the best performance in terms of recall (84%). Including the invasive features in the classification improves the performance significantly with both recall and precision increasing by more than 20 percentage points. We were able to predict the MAP with a root mean square error (a frequently used measure of the differences between the predicted values and the observed values) of 10 mmHg 60 minutes in the future. After converting continuous MAP predictions into AHE binary predictions, we achieve a 91% recall and 68% precision. In addition to predicting AHE, the MAP predictions provide clinically useful information regarding the timing and severity of the AHE occurrence. Conclusion We were able to predict AHE with precision and recall above 80% 30 minutes in advance with the large real-world dataset. The prediction of regression model can provide a more fine-grained, interpretable signal to practitioners. Model performance is improved by the inclusion of invasive features in predicting AHE, when compared to predicting the AHE based on only the available, restricted set of noninvasive technologies. This demonstrates the importance of exploring more noninvasive technologies for AHE prediction.


Author(s):  
Paul Oehlmann ◽  
Paul Osswald ◽  
Juan Camilo Blanco ◽  
Martin Friedrich ◽  
Dominik Rietzel ◽  
...  

AbstractWith industries pushing towards digitalized production, adaption to expectations and increasing requirements for modern applications, has brought additive manufacturing (AM) to the forefront of Industry 4.0. In fact, AM is a main accelerator for digital production with its possibilities in structural design, such as topology optimization, production flexibility, customization, product development, to name a few. Fused Filament Fabrication (FFF) is a widespread and practical tool for rapid prototyping that also demonstrates the importance of AM technologies through its accessibility to the general public by creating cost effective desktop solutions. An increasing integration of systems in an intelligent production environment also enables the generation of large-scale data to be used for process monitoring and process control. Deep learning as a form of artificial intelligence (AI) and more specifically, a method of machine learning (ML) is ideal for handling big data. This study uses a trained artificial neural network (ANN) model as a digital shadow to predict the force within the nozzle of an FFF printer using filament speed and nozzle temperatures as input data. After the ANN model was tested using data from a theoretical model it was implemented to predict the behavior using real-time printer data. For this purpose, an FFF printer was equipped with sensors that collect real time printer data during the printing process. The ANN model reflected the kinematics of melting and flow predicted by models currently available for various speeds of printing. The model allows for a deeper understanding of the influencing process parameters which ultimately results in the determination of the optimum combination of process speed and print quality.


Author(s):  
Petar Radanliev ◽  
David De Roure ◽  
Kevin Page ◽  
Max Van Kleek ◽  
Omar Santos ◽  
...  

AbstractMultiple governmental agencies and private organisations have made commitments for the colonisation of Mars. Such colonisation requires complex systems and infrastructure that could be very costly to repair or replace in cases of cyber-attacks. This paper surveys deep learning algorithms, IoT cyber security and risk models, and established mathematical formulas to identify the best approach for developing a dynamic and self-adapting system for predictive cyber risk analytics supported with Artificial Intelligence and Machine Learning and real-time intelligence in edge computing. The paper presents a new mathematical approach for integrating concepts for cognition engine design, edge computing and Artificial Intelligence and Machine Learning to automate anomaly detection. This engine instigates a step change by applying Artificial Intelligence and Machine Learning embedded at the edge of IoT networks, to deliver safe and functional real-time intelligence for predictive cyber risk analytics. This will enhance capacities for risk analytics and assists in the creation of a comprehensive and systematic understanding of the opportunities and threats that arise when edge computing nodes are deployed, and when Artificial Intelligence and Machine Learning technologies are migrated to the periphery of the internet and into local IoT networks.


Energies ◽  
2021 ◽  
Vol 14 (4) ◽  
pp. 1055
Author(s):  
Qian Sun ◽  
William Ampomah ◽  
Junyu You ◽  
Martha Cather ◽  
Robert Balch

Machine-learning technologies have exhibited robust competences in solving many petroleum engineering problems. The accurate predictivity and fast computational speed enable a large volume of time-consuming engineering processes such as history-matching and field development optimization. The Southwest Regional Partnership on Carbon Sequestration (SWP) project desires rigorous history-matching and multi-objective optimization processes, which fits the superiorities of the machine-learning approaches. Although the machine-learning proxy models are trained and validated before imposing to solve practical problems, the error margin would essentially introduce uncertainties to the results. In this paper, a hybrid numerical machine-learning workflow solving various optimization problems is presented. By coupling the expert machine-learning proxies with a global optimizer, the workflow successfully solves the history-matching and CO2 water alternative gas (WAG) design problem with low computational overheads. The history-matching work considers the heterogeneities of multiphase relative characteristics, and the CO2-WAG injection design takes multiple techno-economic objective functions into accounts. This work trained an expert response surface, a support vector machine, and a multi-layer neural network as proxy models to effectively learn the high-dimensional nonlinear data structure. The proposed workflow suggests revisiting the high-fidelity numerical simulator for validation purposes. The experience gained from this work would provide valuable guiding insights to similar CO2 enhanced oil recovery (EOR) projects.


Sign in / Sign up

Export Citation Format

Share Document