Enhanced Mineral Quantification and Uncertainty Analysis from Downhole Spectroscopy Logs Using Variational Autoencoders

Author(s):  
Paul R. Craddock ◽  
◽  
Prakhar Srivastava ◽  
Harish Datir ◽  
David Rose ◽  
...  

This paper describes an innovative machine-learning application, based on variational autoencoder frameworks, to quantify the concentrations and associated uncertainties of common minerals in sedimentary formations using the measurement of atomic element concentrations from geochemical spectroscopy logs as inputs. The algorithm comprises an input(s), encoder, decoder, output(s), and a novel cost function to optimize the model coefficients during training. The input to the algorithm is a set of dry-weight concentrations of atomic elements with their associated uncertainty. The first output is a set of dry-weight fractions of 14 minerals, and the second output is a set of reconstructed dry-weight concentrations of the original elements. Both sets of outputs include estimates of uncertainty on their predictions. The encoder and decoder are multilayer feed-forward artificial neural networks (ANN), with their coefficients (weights) optimized during calibration (training). The cost function simultaneously minimizes error (accuracy metric) and variance (precision or robustness metric) on the mineral and reconstructed elemental outputs. Training of the weights is done using a set of several-thousand core samples with independent, high-fidelity elemental and mineral (quartz, potassium-feldspar, plagioclase-feldspar, illite, smectite, kaolinite, chlorite, mica, calcite, dolomite, ankerite, siderite, pyrite, and anhydrite) data. The algorithm provides notable advantages over existing methods to estimate formation lithology or mineralogy relying on simple linear, empirical, or nearest-neighbor functions. The ANN numerically capture the multidimensional and nonlinear geochemical relationship (mapping) between elements and minerals that is insufficiently described by prior methods. Training is iterative via backpropagation and samples from Gaussian distributions on each of the elemental inputs, rather than single values, for every sample at each iteration (epoch). These Gaussian distributions are chosen to specifically represent the unique statistical uncertainty of the dry-weight elements in the logging measurements. Sampling from Gaussian distributions during training reduces the potential for overfitting, provides robustness for log interpretations, and further enables a calibrated estimate of uncertainty on the mineral and reconstructed elemental outputs, all of which are lacking in prior methods. The framework of the algorithm is purposefully generalizable so that it can be adapted across geochemical spectroscopy tools. The algorithm reasonably approximates a “global-average” model that requires neither different calibrations nor expert parameterization or intervention for interpreting common oilfield sedimentary formations, although the framework is again purposefully generalizable so it can be optimized for local environments where desirable. The paper showcases a field application of the method for estimating mineral type and abundance in oilfield formations from wellbore-logging measurements.

2021 ◽  
Author(s):  
Paul Craddock ◽  
◽  
Prakhar Srivastava ◽  
Harish Datir ◽  
David Rose ◽  
...  

This paper describes an innovative machine learning application, based on variational autoencoder frameworks, to quantify the concentrations and associated uncertainties of common minerals in sedimentary formations using the measurement of atomic element concentrations from geochemical spectroscopy logs as inputs. The algorithm comprises an input(s), encoder, decoder, output(s), and a novel cost function to optimize the model coefficients during training. The input to the algorithm is a set of dry-weight concentrations of atomic elements with their associated uncertainty. The first output is a set of dry-weight fractions of fourteen minerals, and the second output is a set of reconstructed dry-weight concentrations of the original elements. Both sets of outputs include estimates of uncertainty on their predictions. The encoder and decoder are multilayer feed-forward artificial neural networks (ANN), with their coefficients (weights) optimized during calibration (training). The cost function simultaneously minimizes error (the accuracy metric) and variance (the precision or robustness metric) on the mineral and reconstructed elemental outputs. Training of the weights is done using a set of several-thousand core samples with independent, high-fidelity elemental and mineral (quartz, potassium-feldspar, plagioclase-feldspar, illite, smectite, kaolinite, chlorite, mica, calcite, dolomite, ankerite, siderite, pyrite, and anhydrite) data. The algorithm provides notable advantages over existing methods to estimate formation lithology or mineralogy relying on simple linear, empirical, or nearest-neighbor functions. The ANN numerically capture the multi-dimensional and nonlinear geochemical relationship (mapping) between elements and minerals that is insufficiently described by prior methods. Training is iterative via backpropagation and samples from Gaussian distributions on each of the elemental inputs, rather than single values, for every sample at each iteration (epoch). These Gaussian distributions are chosen to specifically represent the unique statistical uncertainty of the dry-weight elements in the logging measurements. Sampling from Gaussian distributions during training reduces the potential for overfitting, provides robustness for log interpretations, and further enables a calibrated estimate of uncertainty on the mineral and reconstructed elemental outputs, all of which are lacking in prior methods. The framework of the algorithm is purposefully generalizable that it can be adapted across geochemical spectroscopy tools. The algorithm reasonably approximates a ‘global-average’ model that requires neither different calibrations nor expert parameterization or intervention for interpreting common oilfield sedimentary formations, although the framework is again purposefully generalizable so it can be optimized for local environments where desirable. The paper showcases field application of the method for estimating mineral type and abundance in oilfield formations from wellbore logging measurements.


2021 ◽  
Vol 11 (2) ◽  
pp. 850
Author(s):  
Dokkyun Yi ◽  
Sangmin Ji ◽  
Jieun Park

Artificial intelligence (AI) is achieved by optimizing the cost function constructed from learning data. Changing the parameters in the cost function is an AI learning process (or AI learning for convenience). If AI learning is well performed, then the value of the cost function is the global minimum. In order to obtain the well-learned AI learning, the parameter should be no change in the value of the cost function at the global minimum. One useful optimization method is the momentum method; however, the momentum method has difficulty stopping the parameter when the value of the cost function satisfies the global minimum (non-stop problem). The proposed method is based on the momentum method. In order to solve the non-stop problem of the momentum method, we use the value of the cost function to our method. Therefore, as the learning method processes, the mechanism in our method reduces the amount of change in the parameter by the effect of the value of the cost function. We verified the method through proof of convergence and numerical experiments with existing methods to ensure that the learning works well.


2020 ◽  
Vol 18 (02) ◽  
pp. 2050006 ◽  
Author(s):  
Alexsandro Oliveira Alexandrino ◽  
Carla Negri Lintzmayer ◽  
Zanoni Dias

One of the main problems in Computational Biology is to find the evolutionary distance among species. In most approaches, such distance only involves rearrangements, which are mutations that alter large pieces of the species’ genome. When we represent genomes as permutations, the problem of transforming one genome into another is equivalent to the problem of Sorting Permutations by Rearrangement Operations. The traditional approach is to consider that any rearrangement has the same probability to happen, and so, the goal is to find a minimum sequence of operations which sorts the permutation. However, studies have shown that some rearrangements are more likely to happen than others, and so a weighted approach is more realistic. In a weighted approach, the goal is to find a sequence which sorts the permutations, such that the cost of that sequence is minimum. This work introduces a new type of cost function, which is related to the amount of fragmentation caused by a rearrangement. We present some results about the lower and upper bounds for the fragmentation-weighted problems and the relation between the unweighted and the fragmentation-weighted approach. Our main results are 2-approximation algorithms for five versions of this problem involving reversals and transpositions. We also give bounds for the diameters concerning these problems and provide an improved approximation factor for simple permutations considering transpositions.


2005 ◽  
Vol 133 (6) ◽  
pp. 1710-1726 ◽  
Author(s):  
Milija Zupanski

Abstract A new ensemble-based data assimilation method, named the maximum likelihood ensemble filter (MLEF), is presented. The analysis solution maximizes the likelihood of the posterior probability distribution, obtained by minimization of a cost function that depends on a general nonlinear observation operator. The MLEF belongs to the class of deterministic ensemble filters, since no perturbed observations are employed. As in variational and ensemble data assimilation methods, the cost function is derived using a Gaussian probability density function framework. Like other ensemble data assimilation algorithms, the MLEF produces an estimate of the analysis uncertainty (e.g., analysis error covariance). In addition to the common use of ensembles in calculation of the forecast error covariance, the ensembles in MLEF are exploited to efficiently calculate the Hessian preconditioning and the gradient of the cost function. A sufficient number of iterative minimization steps is 2–3, because of superior Hessian preconditioning. The MLEF method is well suited for use with highly nonlinear observation operators, for a small additional computational cost of minimization. The consistent treatment of nonlinear observation operators through optimization is an advantage of the MLEF over other ensemble data assimilation algorithms. The cost of MLEF is comparable to the cost of existing ensemble Kalman filter algorithms. The method is directly applicable to most complex forecast models and observation operators. In this paper, the MLEF method is applied to data assimilation with the one-dimensional Korteweg–de Vries–Burgers equation. The tested observation operator is quadratic, in order to make the assimilation problem more challenging. The results illustrate the stability of the MLEF performance, as well as the benefit of the cost function minimization. The improvement is noted in terms of the rms error, as well as the analysis error covariance. The statistics of innovation vectors (observation minus forecast) also indicate a stable performance of the MLEF algorithm. Additional experiments suggest the amplified benefit of targeted observations in ensemble data assimilation.


2000 ◽  
Vol 25 (2) ◽  
pp. 209-227 ◽  
Author(s):  
Keith R. McLaren ◽  
Peter D. Rossitter ◽  
Alan A. Powell

1996 ◽  
Vol 465 ◽  
Author(s):  
R. D. Rogers ◽  
M. A. Hamilton ◽  
L. O. Nelson ◽  
J. Benson ◽  
M. Green

ABSTRACTBecause there are literally square kilometers of radioactively contaminated concrete surfaces within the U.S. Department of Energy (DOE) complex, the task (both scope and cost) of decontamination is staggering. Complex-wide cleanup using conventional methodology does not appear to be feasible for every facility because of prioritization, cost, and manual effort required.We are investigating the feasibility of using microbially influenced degradation (MID) of concrete as a unique, innovative approach for the decontamination of concrete. Currently, work is being conducted to determine the practicality and cost effectiveness of using this environmentally acceptable method for decontamination of large surface concrete structures. Under laboratory conditions, the biodecontamination process has successfully been used to remove 2 mm of the surface of concrete slabs. Subsequently, initial field application data from an ongoing pilot-scale demonstration have shown that an average of 2 mm of surface can be removed from meter-square areas of contaminated concrete. The cost for the process has been estimated as $1.29/m2. Methodologies for field application of the process are being developed and will be tested. This paper provides information on the MID process, laboratory evaluation of its use for decontamination, and results from the pilot field application.


2021 ◽  
pp. 107754632110324
Author(s):  
Berk Altıner ◽  
Bilal Erol ◽  
Akın Delibaşı

Adaptive optics systems are powerful tools that are implemented to degrade the effects of wavefront aberrations. In this article, the optimal actuator placement problem is addressed for the improvement of disturbance attenuation capability of adaptive optics systems due to the fact that actuator placement is directly related to the enhancement of system performance. For this purpose, the linear-quadratic cost function is chosen, so that optimized actuator layouts can be specialized according to the type of wavefront aberrations. It is then considered as a convex optimization problem, and the cost function is formulated for the disturbance attenuation case. The success of the presented method is demonstrated by simulation results.


2014 ◽  
Vol 665 ◽  
pp. 643-646
Author(s):  
Ying Liu ◽  
Yan Ye ◽  
Chun Guang Li

Metalearning algorithm learns the base learning algorithm, targeted for improving the performance of the learning system. The incremental delta-bar-delta (IDBD) algorithm is such a metalearning algorithm. On the other hand, sparse algorithms are gaining popularity due to their good performance and wide applications. In this paper, we propose a sparse IDBD algorithm by taking the sparsity of the systems into account. Thenorm penalty is contained in the cost function of the standard IDBD, which is equivalent to adding a zero attractor in the iterations, thus can speed up convergence if the system of interest is indeed sparse. Simulations demonstrate that the proposed algorithm is superior to the competing algorithms in sparse system identification.


Geophysics ◽  
2002 ◽  
Vol 67 (4) ◽  
pp. 1202-1212 ◽  
Author(s):  
Hervé Chauris ◽  
Mark S. Noble ◽  
Gilles Lambaré ◽  
Pascal Podvin

We present a new method based on migration velocity analysis (MVA) to estimate 2‐D velocity models from seismic reflection data with no assumption on reflector geometry or the background velocity field. Classical approaches using picking on common image gathers (CIGs) must consider continuous events over the whole panel. This interpretive step may be difficult—particularly for applications on real data sets. We propose to overcome the limiting factor by considering locally coherent events. A locally coherent event can be defined whenever the imaged reflectivity locally shows lateral coherency at some location in the image cube. In the prestack depth‐migrated volume obtained for an a priori velocity model, locally coherent events are picked automatically, without interpretation, and are characterized by their positions and slopes (tangent to the event). Even a single locally coherent event has information on the unknown velocity model, carried by the value of the slope measured in the CIG. The velocity is estimated by minimizing these slopes. We first introduce the cost function and explain its physical meaning. The theoretical developments lead to two equivalent expressions of the cost function: one formulated in the depth‐migrated domain on locally coherent events in CIGs and the other in the time domain. We thus establish direct links between different methods devoted to velocity estimation: migration velocity analysis using locally coherent events and slope tomography. We finally explain how to compute the gradient of the cost function using paraxial ray tracing to update the velocity model. Our method provides smooth, inverted velocity models consistent with Kirchhoff‐type migration schemes and requires neither the introduction of interfaces nor the interpretation of continuous events. As for most automatic velocity analysis methods, careful preprocessing must be applied to remove coherent noise such as multiples.


Sign in / Sign up

Export Citation Format

Share Document