A machine learning based monitoring framework for CO2 storage

Author(s):  
Anouar Romdhane ◽  
Scott Bunting ◽  
Jo Eidsvik ◽  
Susan Anyosa ◽  
Per Bergmo

<p>With increasingly visible effects of climate changes and a growing awareness of the possible consequences, Carbon Capture and Storage (CCS) technologies are gaining momentum. Currently preparations are being done in Norway for a full-scale CCS project where CO<sub>2</sub> will be stored in a deep saline aquifer. A possible candidate for such storage is Smeaheia, located in the North Sea.</p><p>One of the main risks related to large scale storage projects is leakage of CO<sub>2</sub> out of the storage complex. It is important to design measurement, monitoring and verification (MMV) plans addressing leakage risk together with other risks related to conformance and containment verification. In general, geophysical monitoring represents a significant part of storage monitoring costs. Tailored and cost- effective geophysical monitoring programs that consider the trade-off between value and cost are therefore required. A risk-based approach can be adopted to plan the monitoring, but another more quantitative approach coming from decision analysis is that of value of information (VOI) analysis. In such an analysis one can define a decision problem and measure the value of information as the additional value obtained by purchasing information before making the decision.</p><p>In this work, we study the VOI of seismic data in a context of CO<sub>2</sub> storage decision making. Our goal is to evaluate when a seismic survey has the highest value when it comes to detecting a potential leakage of CO<sub>2</sub>, in a dynamic decision problem where we can either stop or continue the injection. We describe the proposed workflow and illustrate it through a constructed case study using a simplified Smeaheia model. We combine Monte Carlo and statistical regression techniques to estimate the VOI at different times. In a first stage, we define the decision problem. We then efficiently generate 10000 possible distributions of CO<sub>2</sub> saturation using a reduced order-based reservoir simulation tool. We consider both leaking and non-leaking scenarios and account for uncertainties in petrophysical properties (porosity and permeability distributions). From the simulated saturations of CO<sub>2</sub>, we derive distributions of geophysical properties and model the corresponding seismic data. We then regress those values on the reference seismic data, to estimate the VOI. We evaluate the use of two machine learning based regression techniques- the k-nearest neighbours' regression with principal components and convolutional neural network (CNN). Both results are compared. We observe that VOI estimates obtained using the k-nearest neighbours' regressions were consistently lower than the estimates obtained using the CNN. Through bootstrapping, we show that the k-nearest neighbours approach produced more stable VOI estimates when compared to the neural networks' method. We analyse possible reasons of the high variability observed with neural networks and suggest means to mitigate them.</p><p><strong>Acknowledgments</strong></p><p>This publication has been produced with support from the NCCS Centre (NFR project number 257579/E20).</p>

Geophysics ◽  
2021 ◽  
pp. 1-43
Author(s):  
Chao Zhang ◽  
Mirko van der Baan

Neural networks hold substantial promise to automate various processing and interpretation tasks. Yet their performance is often sub-optimal compared with standard but more closely guided approaches. Lack of performance is often attributed to poor generalization, in particular if fewer training examples are provided than free parameters exist in the machine learning algorithm. In this case the training data are typically memorized instead of the algorithm learning the underlying general trends. Network generalization is improved if the provided samples are representative, in that they describe all features of interest well. We argue that a more subtle condition preventing poor performance is that the provided examples must also be complete; the examples must span the full solution space. Ensuring completeness during training is challenging unless the target application is well understood. We illustrate that one possible solution is to make the problem more general if this greatly increases the number of available training data. For instance, if seismic images are treated as a subclass of natural images, then a deep-learning-based denoiser for seismic data can be trained using exclusively natural images. The latter are widely available. The resulting denoising algorithm has never seen any seismic data during the training stage; yet it displays a performance comparable to standard and advanced random-noise reduction methods. We exclude any seismic data during training to demonstrate the natural images are both complete and representative for this specific task. Furthermore, we apply a novel approach to increase the amount of training data known as double noise injection, providing both noisy input and output images during the training process. Given the importance of network generalization, we hope that insights gained in this study may help improve the performance of a range of machine learning applications in geophysics.


2020 ◽  
Vol 143 (1) ◽  
Author(s):  
Samir F. Jreij ◽  
Whitney J. Trainor-Guitton ◽  
Michael Morphew ◽  
Ivan Lim Chen Ning

Abstract Faults play an important role in recharging many geothermal reservoirs, and seismic information can image the locations of these faults. The value of information (VOI) metric is used to objectively quantify and compare the value of two types of seismic receiver data via a machine learning approach. The demonstrated VOI methodology is novel by including spatial models from seismic data and obtaining the information statistics from machine learning. Our two-dimensional numerical experiments compare images created from sparsely spaced (80 m), two-component geophone sampling to high spatial resolution (1 m), single-component DAS. We used a three-fold cross validation of a U-Net convolutional neural networks to achieve average classification statistics. The results suggest that when horizontal sources are utilized, geophones and DAS identify reflectors and non-reflectors at roughly the same rate. The average F1 score for horizontal DAS is 0.939 and 0.931 for geophones. For images created from a vertical source, DAS performed marginally better (F1 = 0.919) than geophones (F1 = 0.877). Our transferrable methodology can provide guidance on which acquisition scenarios can improve images of important structures in the subsurface and present an efficient method for obtaining reliability statistics from high-dimensional, spatial data.


Sensors ◽  
2021 ◽  
Vol 21 (18) ◽  
pp. 6105
Author(s):  
Naveed Iqbal ◽  
Abdulmajid Lawal ◽  
Azzedine Zerguine

The traditional cable-based geophone network is an inefficient way of seismic data transmission owing to the related cost and weight. The future of oil and gas exploration technology demands large-scale seismic acquisition, versatility, flexibility, scalability, and automation. On the one hand, a typical seismic survey can pile up a massive amount of raw seismic data per day. On the other hand, the need for wireless seismic data transmission remains immense. Moving from pre-wired to wireless geophones faces major challenges given the enormous amount of data that needs to be transmitted from geophones to the on-site data collection center. The most important factor that has been ignored in the previous studies for the realization of wireless seismic data transmission is wireless channel effects. While transmitting the seismic data wirelessly, impairments like interference, multi-path fading, and channel noise need to be considered. Therefore, in this work, a novel amalgamation of blind channel identification and deep neural networks is proposed. As a geophone already is responsible for transmitting a tremendous amount of data under tight timing constraints, the proposed setup eschews sending any additional training signals for the purpose of mitigating the channel effects. Note that the deep neural network is trained only on synthetic seismic data without the need to use real data in the training process. Experiments show that the proposed method gives promising results when applied to the real/field data set.


Author(s):  
Shafagat Mahmudova

The study machine learning for software based on Soft Computing technology. It analyzes Soft Computing components. Their use in software, their advantages and challenges are studied. Machine learning and its features are highlighted. The functions and features of neural networks are clarified, and recommendations were given.


2020 ◽  
Author(s):  
Jingbai Li ◽  
Patrick Reiser ◽  
André Eberhard ◽  
Pascal Friederich ◽  
Steven Lopez

<p>Photochemical reactions are being increasingly used to construct complex molecular architectures with mild and straightforward reaction conditions. Computational techniques are increasingly important to understand the reactivities and chemoselectivities of photochemical isomerization reactions because they offer molecular bonding information along the excited-state(s) of photodynamics. These photodynamics simulations are resource-intensive and are typically limited to 1–10 picoseconds and 1,000 trajectories due to high computational cost. Most organic photochemical reactions have excited-state lifetimes exceeding 1 picosecond, which places them outside possible computational studies. Westermeyr <i>et al.</i> demonstrated that a machine learning approach could significantly lengthen photodynamics simulation times for a model system, methylenimmonium cation (CH<sub>2</sub>NH<sub>2</sub><sup>+</sup>).</p><p>We have developed a Python-based code, Python Rapid Artificial Intelligence <i>Ab Initio</i> Molecular Dynamics (PyRAI<sup>2</sup>MD), to accomplish the unprecedented 10 ns <i>cis-trans</i> photodynamics of <i>trans</i>-hexafluoro-2-butene (CF<sub>3</sub>–CH=CH–CF<sub>3</sub>) in 3.5 days. The same simulation would take approximately 58 years with ground-truth multiconfigurational dynamics. We proposed an innovative scheme combining Wigner sampling, geometrical interpolations, and short-time quantum chemical trajectories to effectively sample the initial data, facilitating the adaptive sampling to generate an informative and data-efficient training set with 6,232 data points. Our neural networks achieved chemical accuracy (mean absolute error of 0.032 eV). Our 4,814 trajectories reproduced the S<sub>1</sub> half-life (60.5 fs), the photochemical product ratio (<i>trans</i>: <i>cis</i> = 2.3: 1), and autonomously discovered a pathway towards a carbene. The neural networks have also shown the capability of generalizing the full potential energy surface with chemically incomplete data (<i>trans</i> → <i>cis</i> but not <i>cis</i> → <i>trans</i> pathways) that may offer future automated photochemical reaction discoveries.</p>


2021 ◽  
Vol 0 (0) ◽  
Author(s):  
Idris Kharroubi ◽  
Thomas Lim ◽  
Xavier Warin

AbstractWe study the approximation of backward stochastic differential equations (BSDEs for short) with a constraint on the gains process. We first discretize the constraint by applying a so-called facelift operator at times of a grid. We show that this discretely constrained BSDE converges to the continuously constrained one as the mesh grid converges to zero. We then focus on the approximation of the discretely constrained BSDE. For that we adopt a machine learning approach. We show that the facelift can be approximated by an optimization problem over a class of neural networks under constraints on the neural network and its derivative. We then derive an algorithm converging to the discretely constrained BSDE as the number of neurons goes to infinity. We end by numerical experiments.


Sensors ◽  
2021 ◽  
Vol 21 (5) ◽  
pp. 1654
Author(s):  
Poojitha Vurtur Badarinath ◽  
Maria Chierichetti ◽  
Fatemeh Davoudi Kakhki

Current maintenance intervals of mechanical systems are scheduled a priori based on the life of the system, resulting in expensive maintenance scheduling, and often undermining the safety of passengers. Going forward, the actual usage of a vehicle will be used to predict stresses in its structure, and therefore, to define a specific maintenance scheduling. Machine learning (ML) algorithms can be used to map a reduced set of data coming from real-time measurements of a structure into a detailed/high-fidelity finite element analysis (FEA) model of the same system. As a result, the FEA-based ML approach will directly estimate the stress distribution over the entire system during operations, thus improving the ability to define ad-hoc, safe, and efficient maintenance procedures. The paper initially presents a review of the current state-of-the-art of ML methods applied to finite elements. A surrogate finite element approach based on ML algorithms is also proposed to estimate the time-varying response of a one-dimensional beam. Several ML regression models, such as decision trees and artificial neural networks, have been developed, and their performance is compared for direct estimation of the stress distribution over a beam structure. The surrogate finite element models based on ML algorithms are able to estimate the response of the beam accurately, with artificial neural networks providing more accurate results.


2021 ◽  
Vol 11 (15) ◽  
pp. 6704
Author(s):  
Jingyong Cai ◽  
Masashi Takemoto ◽  
Yuming Qiu ◽  
Hironori Nakajo

Despite being heavily used in the training of deep neural networks (DNNs), multipliers are resource-intensive and insufficient in many different scenarios. Previous discoveries have revealed the superiority when activation functions, such as the sigmoid, are calculated by shift-and-add operations, although they fail to remove multiplications in training altogether. In this paper, we propose an innovative approach that can convert all multiplications in the forward and backward inferences of DNNs into shift-and-add operations. Because the model parameters and backpropagated errors of a large DNN model are typically clustered around zero, these values can be approximated by their sine values. Multiplications between the weights and error signals are transferred to multiplications of their sine values, which are replaceable with simpler operations with the help of the product to sum formula. In addition, a rectified sine activation function is utilized for further converting layer inputs into sine values. In this way, the original multiplication-intensive operations can be computed through simple add-and-shift operations. This trigonometric approximation method provides an efficient training and inference alternative for devices with insufficient hardware multipliers. Experimental results demonstrate that this method is able to obtain a performance close to that of classical training algorithms. The approach we propose sheds new light on future hardware customization research for machine learning.


SLEEP ◽  
2021 ◽  
Vol 44 (Supplement_2) ◽  
pp. A164-A164
Author(s):  
Pahnwat Taweesedt ◽  
JungYoon Kim ◽  
Jaehyun Park ◽  
Jangwoon Park ◽  
Munish Sharma ◽  
...  

Abstract Introduction Obstructive sleep apnea (OSA) is a common sleep-related breathing disorder with an estimation of one billion people. Full-night polysomnography is considered the gold standard for OSA diagnosis. However, it is time-consuming, expensive and is not readily available in many parts of the world. Many screening questionnaires and scores have been proposed for OSA prediction with high sensitivity and low specificity. The present study is intended to develop models with various machine learning techniques to predict the severity of OSA by incorporating features from multiple questionnaires. Methods Subjects who underwent full-night polysomnography in Torr sleep center, Texas and completed 5 OSA screening questionnaires/scores were included. OSA was diagnosed by using Apnea-Hypopnea Index ≥ 5. We trained five different machine learning models including Deep Neural Networks with the scaled principal component analysis (DNN-PCA), Random Forest (RF), Adaptive Boosting classifier (ABC), and K-Nearest Neighbors classifier (KNC) and Support Vector Machine Classifier (SVMC). Training:Testing subject ratio of 65:35 was used. All features including demographic data, body measurement, snoring and sleepiness history were obtained from 5 OSA screening questionnaires/scores (STOP-BANG questionnaires, Berlin questionnaires, NoSAS score, NAMES score and No-Apnea score). Performance parametrics were used to compare between machine learning models. Results Of 180 subjects, 51.5 % of subjects were male with mean (SD) age of 53.6 (15.1). One hundred and nineteen subjects were diagnosed with OSA. Area Under the Receiver Operating Characteristic Curve (AUROC) of DNN-PCA, RF, ABC, KNC, SVMC, STOP-BANG questionnaire, Berlin questionnaire, NoSAS score, NAMES score, and No-Apnea score were 0.85, 0.68, 0.52, 0.74, 0.75, 0.61, 0.63, 0,61, 0.58 and 0,58 respectively. DNN-PCA showed the highest AUROC with sensitivity of 0.79, specificity of 0.67, positive-predictivity of 0.93, F1 score of 0.86, and accuracy of 0.77. Conclusion Our result showed that DNN-PCA outperforms OSA screening questionnaires, scores and other machine learning models. Support (if any):


Sign in / Sign up

Export Citation Format

Share Document