Machine Learning-Based Improved Pressure–Volume–Temperature Correlations for Black Oil Reservoirs

2021 ◽  
Vol 143 (11) ◽  
Author(s):  
Zeeshan Tariq ◽  
Mohamed Mahmoud ◽  
Abdulazeez Abdulraheem

Abstract Pressure–volume–temperature (PVT) properties of crude oil are considered the most important properties in petroleum engineering applications as they are virtually used in every reservoir and production engineering calculation. Determination of these properties in the laboratory is the most accurate way to obtain a representative value, at the same time, it is very expensive. However, in the absence of such facilities, other approaches such as analytical solutions and empirical correlations are used to estimate the PVT properties. This study demonstrates the combined use of two machine learning (ML) technique, viz., functional network (FN) coupled with particle swarm optimization (PSO) in predicting the black oil PVT properties such as bubble point pressure (Pb), oil formation volume factor at Pb, and oil viscosity at Pb. This study also proposes new mathematical models derived from the coupled FN-PSO model to estimate these properties. The use of proposed mathematical models does not need any ML engine for the execution. A total of 760 data points collected from the different sources were preprocessed and utilized to build and train the machine learning models. The data utilized covered a wide range of values that are quite reasonable in petroleum engineering applications. The performances of the developed models were tested against the most used empirical correlations. The results showed that the proposed PVT models outperformed previous models by demonstrating an error of up to 2%. The proposed FN-PSO models were also compared with other ML techniques such as an artificial neural network, support vector regression, and adaptive neuro-fuzzy inference system, and the results showed that proposed FN-PSO models outperformed other ML techniques.

2021 ◽  
Author(s):  
Abderraouf Chemmakh

Abstract Uniaxial Compressive Strength (UCS) and Tensile Strength (TS) are among the essential rock parameters required and determined for rock mechanical studies in Petroleum Engineering. However, the determination of such parameters requires some laboratory experiments, which may be time-consuming and costly at the same time. In order to estimate these parameters efficiently and in a short period, some mathematical tools have been used by different researchers. When regression tools proved to give good results only in the limited range of data used, machine learning methods proved to be very accurate in generating models that can cover a wide range of data. In this study, two machine learning models were used to predict the UCS and TS, Support Vector Regression optimized by Genetic Algorithm (GA-SVR) and Artificial Neural Networks (ANNs). The results were discussed for both uniaxial compressive strength and tensile strength in terms of coefficient of determination R2, root mean squared error (RMSE) and mean average error (MAE). First, for the case of UCS, values of 0.99 and 0.99, values of 3.41 and 2.9 and values of 2.43 and 1.9 were obtained for R2, RMSE and MAE for the ANN and GA-SVR, respectively. Second, for the TS, the same analogy was followed, a coefficient R2 of 0.99 and 0.99, RMSE values of 0.41 and 0.45 and MAE values of 0.30 and 0.39 were obtained for ANNs and GA-SVR, respectively. The next step was to assess these models on a different dataset consisting of data obtained from Bakken Field in Williston Basin, North Dakota, United States. The models showed excellent results comparing to the correlations they were compared with, outperforming them in terms of R2, RMSE and MAE, giving the following results for ANN and SVR respectively, R2 of 0.93, 0.92, RMSE of 9.54, 11.22 and MAE of 7.28, 9.24. The resultant conclusion of this work is that the use of machine learning algorithms can generate universal models which reduce the time and effort to estimate some complex parameters such as UCS and Tensile Strength.


Energies ◽  
2021 ◽  
Vol 14 (4) ◽  
pp. 1055
Author(s):  
Qian Sun ◽  
William Ampomah ◽  
Junyu You ◽  
Martha Cather ◽  
Robert Balch

Machine-learning technologies have exhibited robust competences in solving many petroleum engineering problems. The accurate predictivity and fast computational speed enable a large volume of time-consuming engineering processes such as history-matching and field development optimization. The Southwest Regional Partnership on Carbon Sequestration (SWP) project desires rigorous history-matching and multi-objective optimization processes, which fits the superiorities of the machine-learning approaches. Although the machine-learning proxy models are trained and validated before imposing to solve practical problems, the error margin would essentially introduce uncertainties to the results. In this paper, a hybrid numerical machine-learning workflow solving various optimization problems is presented. By coupling the expert machine-learning proxies with a global optimizer, the workflow successfully solves the history-matching and CO2 water alternative gas (WAG) design problem with low computational overheads. The history-matching work considers the heterogeneities of multiphase relative characteristics, and the CO2-WAG injection design takes multiple techno-economic objective functions into accounts. This work trained an expert response surface, a support vector machine, and a multi-layer neural network as proxy models to effectively learn the high-dimensional nonlinear data structure. The proposed workflow suggests revisiting the high-fidelity numerical simulator for validation purposes. The experience gained from this work would provide valuable guiding insights to similar CO2 enhanced oil recovery (EOR) projects.


Energies ◽  
2021 ◽  
Vol 14 (4) ◽  
pp. 930
Author(s):  
Fahimeh Hadavimoghaddam ◽  
Mehdi Ostadhassan ◽  
Ehsan Heidaryan ◽  
Mohammad Ali Sadri ◽  
Inna Chapanova ◽  
...  

Dead oil viscosity is a critical parameter to solve numerous reservoir engineering problems and one of the most unreliable properties to predict with classical black oil correlations. Determination of dead oil viscosity by experiments is expensive and time-consuming, which means developing an accurate and quick prediction model is required. This paper implements six machine learning models: random forest (RF), lightgbm, XGBoost, multilayer perceptron (MLP) neural network, stochastic real-valued (SRV) and SuperLearner to predict dead oil viscosity. More than 2000 pressure–volume–temperature (PVT) data were used for developing and testing these models. A huge range of viscosity data were used, from light intermediate to heavy oil. In this study, we give insight into the performance of different functional forms that have been used in the literature to formulate dead oil viscosity. The results show that the functional form f(γAPI,T), has the best performance, and additional correlating parameters might be unnecessary. Furthermore, SuperLearner outperformed other machine learning (ML) algorithms as well as common correlations that are based on the metric analysis. The SuperLearner model can potentially replace the empirical models for viscosity predictions on a wide range of viscosities (any oil type). Ultimately, the proposed model is capable of simulating the true physical trend of the dead oil viscosity with variations of oil API gravity, temperature and shear rate.


2021 ◽  
Vol 2021 ◽  
pp. 1-9
Author(s):  
Osama Siddig ◽  
Ahmed Farid Ibrahim ◽  
Salaheldin Elkatatny

Unconventional resources have recently gained a lot of attention, and as a consequence, there has been an increase in research interest in predicting total organic carbon (TOC) as a crucial quality indicator. TOC is commonly measured experimentally; however, due to sampling restrictions, obtaining continuous data on TOC is difficult. Therefore, different empirical correlations for TOC have been presented. However, there are concerns about the generalization and accuracy of these correlations. In this paper, different machine learning (ML) techniques were utilized to develop models that predict TOC from well logs, including formation resistivity (FR), spontaneous potential (SP), sonic transit time (Δt), bulk density (RHOB), neutron porosity (CNP), gamma ray (GR), and spectrum logs of thorium (Th), uranium (Ur), and potassium (K). Over 1250 data points from the Devonian Duvernay shale were utilized to create and validate the model. These datasets were obtained from three wells; the first was used to train the models, while the data sets from the other two wells were utilized to test and validate them. Support vector machine (SVM), random forest (RF), and decision tree (DT) were the ML approaches tested, and their predictions were contrasted with three empirical correlations. Various AI methods’ parameters were tested to assure the best possible accuracy in terms of correlation coefficient (R) and average absolute percentage error (AAPE) between the actual and predicted TOC. The three ML methods yielded good matches; however, the RF-based model has the best performance. The RF model was able to predict the TOC for the different datasets with R values range between 0.93 and 0.99 and AAPE values less than 14%. In terms of average error, the ML-based models outperformed the other three empirical correlations. This study shows the capability and robustness of ML models to predict the total organic carbon from readily available logging data without the need for core analysis or additional well interventions.


Sensors ◽  
2020 ◽  
Vol 20 (11) ◽  
pp. 3144 ◽  
Author(s):  
Sherif Said ◽  
Ilyes Boulkaibet ◽  
Murtaza Sheikh ◽  
Abdullah S. Karar ◽  
Samer Alkork ◽  
...  

In this paper, a customizable wearable 3D-printed bionic arm is designed, fabricated, and optimized for a right arm amputee. An experimental test has been conducted for the user, where control of the artificial bionic hand is accomplished successfully using surface electromyography (sEMG) signals acquired by a multi-channel wearable armband. The 3D-printed bionic arm was designed for the low cost of 295 USD, and was lightweight at 428 g. To facilitate a generic control of the bionic arm, sEMG data were collected for a set of gestures (fist, spread fingers, wave-in, wave-out) from a wide range of participants. The collected data were processed and features related to the gestures were extracted for the purpose of training a classifier. In this study, several classifiers based on neural networks, support vector machine, and decision trees were constructed, trained, and statistically compared. The support vector machine classifier was found to exhibit an 89.93% success rate. Real-time testing of the bionic arm with the optimum classifier is demonstrated.


DYNA ◽  
2019 ◽  
Vol 86 (211) ◽  
pp. 32-41 ◽  
Author(s):  
Juan D. Pineda-Jaramillo

In recent decades, transportation planning researchers have used diverse types of machine learning (ML) algorithms to research a wide range of topics. This review paper starts with a brief explanation of some ML algorithms commonly used for transportation research, specifically Artificial Neural Networks (ANN), Decision Trees (DT), Support Vector Machines (SVM) and Cluster Analysis (CA). Then, these different methodologies used by researchers for modeling travel mode choice are collected and compared with the Multinomial Logit Model (MNL) which is the most commonly-used discrete choice model. Finally, the characterization of ML algorithms is discussed and Random Forest (RF), a variant of Decision Tree algorithms, is presented as the best methodology for modeling travel mode choice.


2019 ◽  
Author(s):  
Adane Tarekegn ◽  
Fulvio Ricceri ◽  
Giuseppe Costa ◽  
Elisa Ferracin ◽  
Mario Giacobini

BACKGROUND Frailty is one of the most critical age-related conditions in older adults. It is often recognized as a syndrome of physiological decline in late life, characterized by a marked vulnerability to adverse health outcomes. A clear operational definition of frailty, however, has not been agreed so far. There is a wide range of studies on the detection of frailty and their association with mortality. Several of these studies have focused on the possible risk factors associated with frailty in the elderly population while predicting who will be at increased risk of frailty is still overlooked in clinical settings. OBJECTIVE The objective of our study was to develop predictive models for frailty conditions in older people using different machine learning methods based on a database of clinical characteristics and socioeconomic factors. METHODS An administrative health database containing 1,095,612 elderly people aged 65 or older with 58 input variables and 6 output variables was used. We first identify and define six problems/outputs as surrogates of frailty. We then resolve the imbalanced nature of the data through resampling process and a comparative study between the different machine learning (ML) algorithms – Artificial neural network (ANN), Genetic programming (GP), Support vector machines (SVM), Random Forest (RF), Logistic regression (LR) and Decision tree (DT) – was carried out. The performance of each model was evaluated using a separate unseen dataset. RESULTS Predicting mortality outcome has shown higher performance with ANN (TPR 0.81, TNR 0.76, accuracy 0.78, F1-score 0.79) and SVM (TPR 0.77, TNR 0.80, accuracy 0.79, F1-score 0.78) than predicting the other outcomes. On average, over the six problems, the DT classifier has shown the lowest accuracy, while other models (GP, LR, RF, ANN, and SVM) performed better. All models have shown lower accuracy in predicting an event of an emergency admission with red code than predicting fracture and disability. In predicting urgent hospitalization, only SVM achieved better performance (TPR 0.75, TNR 0.77, accuracy 0.73, F1-score 0.76) with the 10-fold cross validation compared with other models in all evaluation metrics. CONCLUSIONS We developed machine learning models for predicting frailty conditions (mortality, urgent hospitalization, disability, fracture, and emergency admission). The results show that the prediction performance of machine learning models significantly varies from problem to problem in terms of different evaluation metrics. Through further improvement, the model that performs better can be used as a base for developing decision-support tools to improve early identification and prediction of frail older adults.


2019 ◽  
Vol 35 (20) ◽  
pp. 4072-4080 ◽  
Author(s):  
Timo M Deist ◽  
Andrew Patti ◽  
Zhaoqi Wang ◽  
David Krane ◽  
Taylor Sorenson ◽  
...  

Abstract Motivation In a predictive modeling setting, if sufficient details of the system behavior are known, one can build and use a simulation for making predictions. When sufficient system details are not known, one typically turns to machine learning, which builds a black-box model of the system using a large dataset of input sample features and outputs. We consider a setting which is between these two extremes: some details of the system mechanics are known but not enough for creating simulations that can be used to make high quality predictions. In this context we propose using approximate simulations to build a kernel for use in kernelized machine learning methods, such as support vector machines. The results of multiple simulations (under various uncertainty scenarios) are used to compute similarity measures between every pair of samples: sample pairs are given a high similarity score if they behave similarly under a wide range of simulation parameters. These similarity values, rather than the original high dimensional feature data, are used to build the kernel. Results We demonstrate and explore the simulation-based kernel (SimKern) concept using four synthetic complex systems—three biologically inspired models and one network flow optimization model. We show that, when the number of training samples is small compared to the number of features, the SimKern approach dominates over no-prior-knowledge methods. This approach should be applicable in all disciplines where predictive models are sought and informative yet approximate simulations are available. Availability and implementation The Python SimKern software, the demonstration models (in MATLAB, R), and the datasets are available at https://github.com/davidcraft/SimKern. Supplementary information Supplementary data are available at Bioinformatics online.


2020 ◽  
Author(s):  
Mazin Mohammed ◽  
Karrar Hameed Abdulkareem ◽  
Mashael S. Maashi ◽  
Salama A. Mostafa A. Mostafa ◽  
Abdullah Baz ◽  
...  

BACKGROUND In most recent times, global concern has been caused by a coronavirus (COVID19), which is considered a global health threat due to its rapid spread across the globe. Machine learning (ML) is a computational method that can be used to automatically learn from experience and improve the accuracy of predictions. OBJECTIVE In this study, the use of machine learning has been applied to Coronavirus dataset of 50 X-ray images to enable the development of directions and detection modalities with risk causes.The dataset contains a wide range of samples of COVID-19 cases alongside SARS, MERS, and ARDS. The experiment was carried out using a total of 50 X-ray images, out of which 25 images were that of positive COVIDE-19 cases, while the other 25 were normal cases. METHODS An orange tool has been used for data manipulation. To be able to classify patients as carriers of Coronavirus and non-Coronavirus carriers, this tool has been employed in developing and analysing seven types of predictive models. Models such as , artificial neural network (ANN), support vector machine (SVM), linear kernel and radial basis function (RBF), k-nearest neighbour (k-NN), Decision Tree (DT), and CN2 rule inducer were used in this study.Furthermore, the standard InceptionV3 model has been used for feature extraction target. RESULTS The various machine learning techniques that have been trained on coronavirus disease 2019 (COVID-19) dataset with improved ML techniques parameters. The data set was divided into two parts, which are training and testing. The model was trained using 70% of the dataset, while the remaining 30% was used to test the model. The results show that the improved SVM achieved a F1 of 97% and an accuracy of 98%. CONCLUSIONS :. In this study, seven models have been developed to aid the detection of coronavirus. In such cases, the learning performance can be improved through knowledge transfer, whereby time-consuming data labelling efforts are not required.the evaluations of all the models are done in terms of different parameters. it can be concluded that all the models performed well, but the SVM demonstrated the best result for accuracy metric. Future work will compare classical approaches with deep learning ones and try to obtain better results. CLINICALTRIAL None


Energies ◽  
2021 ◽  
Vol 14 (5) ◽  
pp. 1263
Author(s):  
Alireza Sarraf Shirazi ◽  
Ian Frigaard

Improving the accuracy of the slurry flow predictions in different operating flow regimes remains a major focus for multiphase flow research, and it is especially targeted at industrial applications such as oil and gas. In this paper we develop a robust integrated method consisting of an artificial neural network (ANN) and support vector regression (SVR) to estimate the critical velocity, the slurry flow regime change, and ultimately, the frictional pressure drop for a solid–liquid slurry flow in a horizontal pipe, covering wide ranges of flow and geometrical parameters. Three distinct datasets were used to develop machine learning models with totals of 100, 325, and 125 data points for critical velocity, and frictional pressure drops for heterogeneous and bed-load regimes respectively. For each dataset, 80% of the data were used for training and the rest 20% for evaluating the out of sample performance. The K-fold technique was used for cross-validation. The prediction results of the developed integrated method showed that it significantly outperforms the widely used existing correlations and models in the literature. Additionally, the proposed integrated method with the average absolute relative error (AARE) of 0.084 outperformed the model developed without regime classification with the AARE of 0.155. The proposed integrated model not only offers reliable predictions over a wide range of operating conditions and different flow regimes for the first time, but also introduces a general framework of how to utilize prior physical knowledge to achieve more reliable performances from machine learning methods.


Sign in / Sign up

Export Citation Format

Share Document