A Proxy Peng-Robinson EOS for Efficient Modeling of Phase Behavior

2021 ◽  
Author(s):  
Mark Zhao ◽  
Ryosuke Okuno

Abstract Equation-of-state (EOS) compositional simulation is commonly used to model the interplay between phase behavior and fluid flow for various reservoir and surface processes. Because of its computational cost, however, there is a critical need for efficient phase-behavior calculations using an EOS. The objective of this research was to develop a proxy model for fugacity coefficient based on the Peng-Robinson EOS for rapid multiphase flash in compositional flow simulation. The proxy model as implemented in this research is to bypass the calculations of fugacity coefficients when the Peng-Robinson EOS has only one root, which is often the case at reservoir conditions. The proxy fugacity model was trained by artificial neural networks (ANN) with over 30 million fugacity coefficients based on the Peng-Robinson EOS. It accurately predicts the Peng- Robinson fugacity coefficient by using four parameters: Am, Bm, Bi, and ΣxiAij. Since these scalar parameters are general, not specific to particular compositions, pressures, and temperatures, the proxy model is applicable to petroleum engineering applications as equally as the original Peng-Robinson EOS. The proxy model is applied to multiphase flash calculations (phase-split and stability), where the cubic equation solutions and fugacity coefficient calculations are bypassed when the Peng-Robinson EOS has one root. The original fugacity coefficient is analytically calculated when the EOS has more than one root, but this occurs only occasionally at reservoir conditions. A case study shows the proxy fugacity model gave a speed-up factor of 3.4% in comparison to the conventional EOS calculation. Case studies also demonstrate accurate multiphase flash results (stability and phase split) and interchangeable proxy models for different fluid cases with different (numbers of) components. This is possible because it predicts the Peng-Robinson fugacity in the variable space that is not specific to composition, temperature, and pressure. For the same reason, non-zero binary iteration parameters do not impair the applicability, accuracy, robustness, and efficiency of the model. As the proxy models are specific to individual components, a combination of proxy models can be used to model for any mixture of components. Tuning of training hyperparameters and training data sampling method helped reduce the mean absolute percent error to less than 0.1% in the ANN modeling. To the best of our knowledge, this is the first generalized proxy model of the Peng-Robinson fugacity that is applicable to any mixture. The proposed model retains the conventional flash iteration, the convergence robustness, and the option of manual parameter tuning for fluid characterization.

2018 ◽  
Author(s):  
Roman Zubatyuk ◽  
Justin S. Smith ◽  
Jerzy Leszczynski ◽  
Olexandr Isayev

<p>Atomic and molecular properties could be evaluated from the fundamental Schrodinger’s equation and therefore represent different modalities of the same quantum phenomena. Here we present AIMNet, a modular and chemically inspired deep neural network potential. We used AIMNet with multitarget training to learn multiple modalities of the state of the atom in a molecular system. The resulting model shows on several benchmark datasets the state-of-the-art accuracy, comparable to the results of orders of magnitude more expensive DFT methods. It can simultaneously predict several atomic and molecular properties without an increase in computational cost. With AIMNet we show a new dimension of transferability: the ability to learn new targets utilizing multimodal information from previous training. The model can learn implicit solvation energy (like SMD) utilizing only a fraction of original training data, and archive MAD error of 1.1 kcal/mol compared to experimental solvation free energies in MNSol database.</p>


Energies ◽  
2021 ◽  
Vol 14 (4) ◽  
pp. 1055
Author(s):  
Qian Sun ◽  
William Ampomah ◽  
Junyu You ◽  
Martha Cather ◽  
Robert Balch

Machine-learning technologies have exhibited robust competences in solving many petroleum engineering problems. The accurate predictivity and fast computational speed enable a large volume of time-consuming engineering processes such as history-matching and field development optimization. The Southwest Regional Partnership on Carbon Sequestration (SWP) project desires rigorous history-matching and multi-objective optimization processes, which fits the superiorities of the machine-learning approaches. Although the machine-learning proxy models are trained and validated before imposing to solve practical problems, the error margin would essentially introduce uncertainties to the results. In this paper, a hybrid numerical machine-learning workflow solving various optimization problems is presented. By coupling the expert machine-learning proxies with a global optimizer, the workflow successfully solves the history-matching and CO2 water alternative gas (WAG) design problem with low computational overheads. The history-matching work considers the heterogeneities of multiphase relative characteristics, and the CO2-WAG injection design takes multiple techno-economic objective functions into accounts. This work trained an expert response surface, a support vector machine, and a multi-layer neural network as proxy models to effectively learn the high-dimensional nonlinear data structure. The proposed workflow suggests revisiting the high-fidelity numerical simulator for validation purposes. The experience gained from this work would provide valuable guiding insights to similar CO2 enhanced oil recovery (EOR) projects.


2013 ◽  
Vol 2013 ◽  
pp. 1-10
Author(s):  
Lei Luo ◽  
Chao Zhang ◽  
Yongrui Qin ◽  
Chunyuan Zhang

With the explosive growth of the data volume in modern applications such as web search and multimedia retrieval, hashing is becoming increasingly important for efficient nearest neighbor (similar item) search. Recently, a number of data-dependent methods have been developed, reflecting the great potential of learning for hashing. Inspired by the classic nonlinear dimensionality reduction algorithm—maximum variance unfolding, we propose a novel unsupervised hashing method, named maximum variance hashing, in this work. The idea is to maximize the total variance of the hash codes while preserving the local structure of the training data. To solve the derived optimization problem, we propose a column generation algorithm, which directly learns the binary-valued hash functions. We then extend it using anchor graphs to reduce the computational cost. Experiments on large-scale image datasets demonstrate that the proposed method outperforms state-of-the-art hashing methods in many cases.


2021 ◽  
Author(s):  
Carlo Cristiano Stabile ◽  
Marco Barbiero ◽  
Giorgio Fighera ◽  
Laura Dovera

Abstract Optimizing well locations for a green field is critical to mitigate development risks. Performing such workflows with reservoir simulations is very challenging due to the huge computational cost. Proxy models can instead provide accurate estimates at a fraction of the computing time. This study presents an application of new generation functional proxies to optimize the well locations in a real oil field with respect to the actualized oil production on all the different geological realizations. Proxies are built with the Universal Trace Kriging and are functional in time allowing to actualize oil flows over the asset lifetime. Proxies are trained on the reservoir simulations using randomly sampled well locations. Two proxies are created for a pessimistic model (P10) and a mid-case model (P50) to capture the geological uncertainties. The optimization step uses the Non-dominated Sorting Genetic Algorithm, with discounted oil productions of the two proxies, as objective functions. An adaptive approach was employed: optimized points found from a first optimization were used to re-train the proxy models and a second run of optimization was performed. The methodology was applied on a real oil reservoir to optimize the location of four vertical production wells and compared against reference locations. 111 geological realizations were available, in which one relevant uncertainty is the presence of possible compartments. The decision space represented by the horizontal translation vectors for each well was sampled using Plackett-Burman and Latin-Hypercube designs. A first application produced a proxy with poor predictive quality. Redrawing the areas to avoid overlaps and to confine the decision space of each well in one compartment, improved the quality. This suggests that the proxy predictive ability deteriorates in presence of highly non-linear responses caused by sealing faults or by well interchanging positions. We then followed a 2-step adaptive approach: a first optimization was performed and the resulting Pareto front was validated with reservoir simulations; to further improve the proxy quality in this region of the decision space, the validated Pareto front points were added to the initial dataset to retrain the proxy and consequently rerun the optimization. The final well locations were validated on all 111 realizations with reservoir simulations and resulted in an overall increase of the discounted production of about 5% compared to the reference development strategy. The adaptive approach, combined with functional proxy, proved to be successful in improving the workflow by purposefully increasing the training set samples with data points able to enhance the optimization step effectiveness. Each optimization run performed relies on about 1 million proxy evaluations which required negligible computational time. The same workflow carried out with standard reservoir simulations would have been practically unfeasible.


SPE Journal ◽  
2021 ◽  
pp. 1-19
Author(s):  
Yingnan Wang ◽  
Nadia Shardt ◽  
Janet A. W. Elliott ◽  
Zhehui Jin

Summary Gas-alkane interfacial tension (IFT) is an important parameter in the enhanced oil recovery (EOR) process. Thus, it is imperative to obtain an accurate gas-alkane mixture IFT for both chemical and petroleum engineering applications. Various empirical correlations have been developed in the past several decades. Although these models are often easy to implement, their accuracy is inconsistent over a wide range of temperatures, pressures, and compositions. Although statistical mechanics-based models and molecular simulations can accurately predict gas-alkane IFT, they usually come with an extensive computational cost. The Shardt-Elliott (SE) model is a highly accurate IFT model that for subcritical fluids is analytic in terms of temperature T and composition x. In applications, it is desirable to obtain IFT in terms of temperature T and pressure P, which requires time-consuming flash calculations, and for mixtures that contain a gas component greater than its pure species critical point, additional critical composition calculations are required. In this work, the SE model is combined with a machine learning (ML) approach to obtain highly efficient and highly accurate gas-alkane binary mixture IFT equations directly in terms of temperature, pressure, and alkane molar weights. The SE model is used to build an IFT database (more than 36,000 points) for ML training to obtain IFT equations. The ML-based IFT equations are evaluated in comparison with the available experimental data (888 points) and with the SE model, as well as with the less accurate parachor model. Overall, the ML-based IFT equations show excellent agreement with experimental data for gas-alkane binary mixtures over a wide range of T and P, and they outperform the widely used parachor model. The developed highly efficient and highly accurate IFT functions can serve as a basis for modeling gas-alkane binary mixtures for a broad range of T, P, and x.


Author(s):  
Tich Phuoc Tran ◽  
Pohsiang Tsai ◽  
Tony Jan ◽  
Xiangjian He

Most of the currently available network security techniques are not able to cope with the dynamic and increasingly complex nature of cyber attacks on distributed computer systems. Therefore, an automated and adaptive defensive tool is imperative for computer networks. Alongside the existing prevention techniques such as encryption and firewalls, Intrusion Detection System (IDS) has established itself as an emerging technology that is able to detect unauthorized access and abuse of computer systems by both internal users and external offenders. Most of the novel approaches in this field have adopted Artificial Intelligence (AI) technologies such as Artificial Neural Networks (ANN) to improve performance as well as robustness of IDS. The true power and advantages of ANN lie in its ability to represent both linear and non-linear relationships and learn these relationships directly from the data being modeled. However, ANN is computationally expensive due to its demanding processing power and this leads to overfitting problem, i.e. the network is unable to extrapolate accurately once the input is outside of the training data range. These limitations challenge IDS with low detection rate, high false alarm rate and excessive computation cost. This chapter proposes a novel Machine Learning (ML) algorithm to alleviate those difficulties of existing AI techniques in the area of computer network security. The Intrusion Detection dataset provided by Knowledge Discovery and Data Mining (KDD-99) is used as a benchmark to compare our model with other existing techniques. Extensive empirical analysis suggests that the proposed method outperforms other state-of-the-art learning algorithms in terms of learning bias, generalization variance and computational cost. It is also reported to significantly improve the overall detection capability for difficult-to-detect novel attacks which are unseen or irregularly occur in the training phase.


2020 ◽  
Vol 10 (2) ◽  
pp. 483 ◽  
Author(s):  
Eko Ihsanto ◽  
Kalamullah Ramli ◽  
Dodi Sudiana ◽  
Teddy Surya Gunawan

Many algorithms have been developed for automated electrocardiogram (ECG) classification. Due to the non-stationary nature of the ECG signal, it is rather challenging to use traditional handcraft methods, such as time-based analysis of feature extraction and classification, to pave the way for machine learning implementation. This paper proposed a novel method, i.e., the ensemble of depthwise separable convolutional (DSC) neural networks for the classification of cardiac arrhythmia ECG beats. Using our proposed method, the four stages of ECG classification, i.e., QRS detection, preprocessing, feature extraction, and classification, were reduced to two steps only, i.e., QRS detection and classification. No preprocessing method was required while feature extraction was combined with classification. Moreover, to reduce the computational cost while maintaining its accuracy, several techniques were implemented, including All Convolutional Network (ACN), Batch Normalization (BN), and ensemble convolutional neural networks. The performance of the proposed ensemble CNNs were evaluated using the MIT-BIH arrythmia database. In the training phase, around 22% of the 110,057 beats data extracted from 48 records were utilized. Using only these 22% labeled training data, our proposed algorithm was able to classify the remaining 78% of the database into 16 classes. Furthermore, the sensitivity ( S n ), specificity ( S p ), and positive predictivity ( P p ), and accuracy ( A c c ) are 99.03%, 99.94%, 99.03%, and 99.88%, respectively. The proposed algorithm required around 180 μs, which is suitable for real time application. These results showed that our proposed method outperformed other state of the art methods.


2020 ◽  
Author(s):  
Hossein Foroozand ◽  
Steven V. Weijs

&lt;p&gt;Machine learning is the fast-growing branch of data-driven models, and its main objective is to use computational methods to become more accurate in predicting outcomes without being explicitly programmed. In this field, a way to improve model predictions is to use a large collection of models (called ensemble) instead of a single one. Each model is then trained on slightly different samples of the original data, and their predictions are averaged. This is called bootstrap aggregating, or Bagging, and is widely applied. A recurring question in previous works was: how to choose the ensemble size of training data sets for tuning the weights in machine learning? The computational cost of ensemble-based methods scales with the size of the ensemble, but excessively reducing the ensemble size comes at the cost of reduced predictive performance. The choice of ensemble size was often determined based on the size of input data and available computational power, which can become a limiting factor for larger datasets and complex models&amp;#8217; training. In this research, it is our hypothesis that if an ensemble of artificial neural networks (ANN) models or any other machine learning technique uses the most informative ensemble members for training purpose rather than all bootstrapped ensemble members, it could reduce the computational time substantially without negatively affecting the performance of simulation.&lt;/p&gt;


2018 ◽  
Vol 30 (4) ◽  
pp. 513-522 ◽  
Author(s):  
Yuichi Konishi ◽  
◽  
Kosuke Shigematsu ◽  
Takashi Tsubouchi ◽  
Akihisa Ohya

The Tsukuba Challenge is an open experiment competition held annually since 2007, and wherein the autonomous navigation robots developed by the participants must navigate through an urban setting in which pedestrians and cyclists are present. One of the required tasks in the Tsukuba Challenge from 2013 to 2017 was to search for persons wearing designated clothes within the search area. This is a very difficult task since it is necessary to seek out these persons in an environment that includes regular pedestrians, and wherein the lighting changes easily because of weather conditions. Moreover, the recognition system must have a light computational cost because of the limited performance of the computer that is mounted onto the robot. In this study, we focused on a deep learning method of detecting the target persons in captured images. The developed detection system was expected to achieve high detection performance, even when small-sized input images were used for deep learning. Experiments demonstrated that the proposed system achieved better performance than an existing object detection network. However, because a vast amount of training data is necessary for deep learning, a method of generating training data to be used in the detection of target persons is also discussed in this paper.


SPE Journal ◽  
2017 ◽  
Vol 23 (03) ◽  
pp. 819-830 ◽  
Author(s):  
V. A. Torrealba ◽  
R. T. Johns

Summary Surfactant-based enhanced oil recovery (EOR) is a promising technique because of surfactant's ability to mobilize previously trapped oil by significantly reducing capillary forces at the pore scale. However, the field-implementation of these techniques is challenged by the high cost of chemicals, which makes the margin of error for the deployment of such methods increasingly narrow. Some commonly recognized issues are surfactant adsorption, surfactant partitioning to the excess phases, thermal and physical degradation, and scale-representative phase behavior. Recent contributions to the petroleum-engineering literature have used the hydrophilic/lipophilic-difference net-average-curvature (HLD-NAC) model to develop a phase-behavior equation of state (EoS) to fit experimental data and predict phase behavior away from tuned data. The model currently assumes spherical micelles and constant three-phase correlation length, which may yield errors in the bicontinuous region where micelles transition into cylindrical and planar shapes. In this paper, we introduce a new empirical phase-behavior model that is based on chemical-potential (CP) trends and HLD that eliminates NAC so that spherical micelles and the constant three-phase correlation length are no longer assumed. The model is able to describe all two-phase regions, and is shown to represent accurately experimental data at fixed composition and changing HLD (e.g., a salinity scan) as well as variable-composition data at fixed HLD. Further, the model is extended to account for surfactant partitioning into the excess phases. The model is benchmarked against experimental data (considering both pure-alkane and crude-oil cases), showing excellent fits and predictions for a wide variety of experiments, and is compared to the recently developed HLD-NAC EoS model for reference.


Sign in / Sign up

Export Citation Format

Share Document