Geologic provinces from unsupervised learning: synthetic experiments in clustering of localized topography/gravity admittance and correlation spectra

Author(s):  
Alberto Pastorutti ◽  
Carla Braitenberg

<p>Partitioning of the Earth surface in "provinces": tectonic domains, outcropping geological units, crustal types, discrete classes extracted from age or geophysical data (e.g. tomography, gravity) is often employed to perform data imputation of ill-sampled observables (e.g. the similarity-based NGHF surface heat flow map [1]) or to constrain the parameters of ill-posed inverse problems (e.g. the gravimetric global Moho model GEMMA [2]).</p><p>We define provinces as noncontiguous areas where quantities or their relationships are similar. Following the goodness metric employed for proxy observables, an adequate province model should be able to significantly improve prediction of the extrapolated quantity. Interpolation of a quantity with no reliance on external data sets a predictivity benchmark, which a province-based prediction should exceed.<br>In a solid Earth modelling perspective, gravity, topography, and their relationship, seem ideal candidates to constrain a province clustering model. Earth gravity and topography, at resolutions of at least 100 km, are known with an incomparable sampling uniformity and negligible error, respect to other observables.</p><p>Most of the observed topography-gravity relationship can be explained by regional isostatic compensation. The topography, representing the load exerted on the lithosphere, is compensated by the elastic, thin-shell like response of the latter. In the spectral domain, flexure results in a lowpass transfer function between topography and isostatic roots. The signal of both surfaces, superimposed, is observed in the gravity field.<br>However, reality shows significant shifts from the ideal case: the separation of nonisostatic effects [3], such as density inhomogeneities, glacial isostatic adjustments, dynamic mantle processes, is nontrivial. Acknowledging this superposition, we aim at identifying clusters of similar topography-gravity transfer functions.</p><p>We evaluate the transfer functions, in the form of admittance and correlation [4], in the spherical harmonics domain. Spatial localization is achieved with the method by Wieczorek and Simons [5], using SHTOOLS [6]. Admittance and correlation spectra are computed on a set of regularly spaced sample points, each point being representative of the topo-gravity relationship in its proximity. The coefficients of the localized topo-gravity admittance and correlation spectra constitute each point features.</p><p>We present a set of experiments performed on synthetic models, in which we can control the variations of elastic parameters and non-isostatic contributions. These tests allowed to define both the feature extraction segment: the spatial localization method and the range of spherical harmonics degrees which are more sensible to lateral variations in flexural rigidity; and the clustering segment: metrics of the ground-truth clusters, performance of dimensionality reduction methods and of different clustering models.</p><p>[1] Lucazeau (2019). Analysis and Mapping of an Updated Terrestrial Heat Flow Data Set. doi:10.1029/2019GC008389<br>[2] Reguzzoni and Sampietro (2015). GEMMA: An Earth crustal model based on GOCE satellite data. doi:10.1016/j.jag.2014.04.002<br>[3] Bagherbandi and Sjöberg (2013). Improving gravimetric–isostatic models of crustal depth by correcting for non-isostatic effects and using CRUST2.0. doi:10.1016/j.earscirev.2012.12.002<br>[4] Simons et al. (1997). Localization of gravity and topography: Constraints on the tectonics and mantle dynamics of Venus. doi:10.1111/j.1365-246X.1997.tb00593.x<br>[5] Wieczorek and Simons (2005). Localized spectral analysis on the sphere. doi:10.1111/j.1365-246X.2005.02687.x<br>[6] Wieczorek and Meschede (2018). SHTools: Tools for Working with Spherical Harmonics. doi:10.1029/2018GC007529</p>

Author(s):  
Sven Fuchs ◽  
Graeme Beardsmore ◽  
Paolo Chiozzi ◽  
Orlando Miguel Espinoza-Ojeda ◽  
Gianluca Gola ◽  
...  

Periodic revisions of the Global Heat Flow Database (GHFD) take place under the auspices of the International Heat Flow Commission (IHFC) of the International Association of Seismology and Physics of the Earth's Interior (IASPEI). A growing number of heat-flow values, advances in scientific methods, digitization, and improvements in database technologies all warrant a revision of the structure of the GHFD that was last amended in 1976. We present a new structure for the GHFD, which will provide a basis for a reassessment and revision of the existing global heat-flow data set. The database fields within the new structure are described in detail to ensure a common understanding of the respective database entries. The new structure of the database takes advantage of today's possibilities for data management. It supports FAIR and open data principles, including interoperability with external data services, and links to DOI and IGSN numbers and other data resources (e.g., world geological map, world stratigraphic system, and International Ocean Drilling Program data). Aligned with this publication, a restructured version of the existing database is published, which provides a starting point for the upcoming collaborative process of data screening, quality control and revision. In parallel, the IHFC will work on criteria for a new quality scheme that will allow future users of the database to evaluate the quality of the collated heat-flow data based on specific criteria.


2019 ◽  
Vol 11 (10) ◽  
pp. 1157 ◽  
Author(s):  
Jorge Fuentes-Pacheco ◽  
Juan Torres-Olivares ◽  
Edgar Roman-Rangel ◽  
Salvador Cervantes ◽  
Porfirio Juarez-Lopez ◽  
...  

Crop segmentation is an important task in Precision Agriculture, where the use of aerial robots with an on-board camera has contributed to the development of new solution alternatives. We address the problem of fig plant segmentation in top-view RGB (Red-Green-Blue) images of a crop grown under open-field difficult circumstances of complex lighting conditions and non-ideal crop maintenance practices defined by local farmers. We present a Convolutional Neural Network (CNN) with an encoder-decoder architecture that classifies each pixel as crop or non-crop using only raw colour images as input. Our approach achieves a mean accuracy of 93.85% despite the complexity of the background and a highly variable visual appearance of the leaves. We make available our CNN code to the research community, as well as the aerial image data set and a hand-made ground truth segmentation with pixel precision to facilitate the comparison among different algorithms.


2020 ◽  
Vol 499 (4) ◽  
pp. 5641-5652
Author(s):  
Georgios Vernardos ◽  
Grigorios Tsagkatakis ◽  
Yannis Pantazis

ABSTRACT Gravitational lensing is a powerful tool for constraining substructure in the mass distribution of galaxies, be it from the presence of dark matter sub-haloes or due to physical mechanisms affecting the baryons throughout galaxy evolution. Such substructure is hard to model and is either ignored by traditional, smooth modelling, approaches, or treated as well-localized massive perturbers. In this work, we propose a deep learning approach to quantify the statistical properties of such perturbations directly from images, where only the extended lensed source features within a mask are considered, without the need of any lens modelling. Our training data consist of mock lensed images assuming perturbing Gaussian Random Fields permeating the smooth overall lens potential, and, for the first time, using images of real galaxies as the lensed source. We employ a novel deep neural network that can handle arbitrary uncertainty intervals associated with the training data set labels as input, provides probability distributions as output, and adopts a composite loss function. The method succeeds not only in accurately estimating the actual parameter values, but also reduces the predicted confidence intervals by 10 per cent in an unsupervised manner, i.e. without having access to the actual ground truth values. Our results are invariant to the inherent degeneracy between mass perturbations in the lens and complex brightness profiles for the source. Hence, we can quantitatively and robustly quantify the smoothness of the mass density of thousands of lenses, including confidence intervals, and provide a consistent ranking for follow-up science.


2021 ◽  
Vol 6 (1) ◽  
Author(s):  
Peter W. Eide ◽  
Seyed H. Moosavi ◽  
Ina A. Eilertsen ◽  
Tuva H. Brunsell ◽  
Jonas Langerud ◽  
...  

AbstractGene expression-based subtypes of colorectal cancer have clinical relevance, but the representativeness of primary tumors and the consensus molecular subtypes (CMS) for metastatic cancers is not well known. We investigated the metastatic heterogeneity of CMS. The best approach to subtype translation was delineated by comparisons of transcriptomic profiles from 317 primary tumors and 295 liver metastases, including multi-metastatic samples from 45 patients and 14 primary-metastasis sets. Associations were validated in an external data set (n = 618). Projection of metastases onto principal components of primary tumors showed that metastases were depleted of CMS1-immune/CMS3-metabolic signals, enriched for CMS4-mesenchymal/stromal signals, and heavily influenced by the microenvironment. The tailored CMS classifier (available in an updated version of the R package CMScaller) therefore implemented an approach to regress out the liver tissue background. The majority of classified metastases were either CMS2 or CMS4. Nonetheless, subtype switching and inter-metastatic CMS heterogeneity were frequent and increased with sampling intensity. Poor-prognostic value of CMS1/3 metastases was consistent in the context of intra-patient tumor heterogeneity.


Geophysics ◽  
2012 ◽  
Vol 77 (4) ◽  
pp. E301-E315 ◽  
Author(s):  
Thomas Kalscheuer ◽  
Juliane Hübert ◽  
Alexey Kuvshinov ◽  
Tobias Lochbühler ◽  
Laust B. Pedersen

Magnetotelluric (MT), radiomagnetotelluric (RMT), and, in particular, controlled-source audiomagnetotelluric (CSAMT) data are often heavily distorted by near-surface inhomogeneities. We developed a novel scheme to invert MT, RMT, and CSAMT data in the form of scalar or tensorial impedances and vertical magnetic transfer functions simultaneously for layer resistivities and electric and magnetic galvanic distortion parameters. The inversion scheme uses smoothness constraints to regularize layer resistivities and either Marquardt-Levenberg damping or the minimum-solution length criterion to regularize distortion parameters. A depth of investigation range is estimated by comparing layered model sections derived from first- and second-order smoothness constraints. Synthetic examples demonstrate that earth models are reconstructed properly for distorted and undistorted tensorial CSAMT data. In the inversion of scalar CSAMT data, such as the determinant impedance or individual tensor elements, the reduced number of transfer functions inevitably leads to increased ambiguity for distortion parameters. As a consequence of this ambiguity for scalar data, distortion parameters often grow over the iterations to unrealistic absolute values when regularized with the Marquardt-Levenberg scheme. Essentially, compensating relationships between terms containing electric and/or magnetic distortion are used in this growth. In a regularization with the minimum solution length criterion, the distortion parameters converge into a stable configuration after several iterations and attain reasonable values. The inversion algorithm was applied to a CSAMT field data set collected along a profile over a tunnel construction site at Hallandsåsen, Sweden. To avoid erroneous inverse models from strong anthropogenic effects on the data, two scalar transfer functions (one scalar impedance and one scalar vertical magnetic transfer function) were selected for inversion. Compared with a regularization of distortion parameters with the Marquardt-Levenberg method, the minimum-solution length criterion yielded smaller absolute values of distortion parameters and a horizontally more homogeneous distribution of electrical conductivity.


Author(s):  
Rohit Shankaran ◽  
Alexander Rimmer ◽  
Alan Haig

In recent years due to use of drilling risers with larger and heavier BOP/LMRP stacks, fatigue loading on subsea wellheads has increased, which poses potential restrictions on the duration of drilling operations. In order to track wellhead and conductor fatigue capacity consumption to support safe drilling operations a range of methods have been applied: • Analytical riser model and measured environmental data; • BOP motion measurement and transfer functions; • Strain gauge data. Strain gauge monitoring is considered the most accurate method for measuring fatigue capacity consumption. To compare the three approaches and establish recommendations for an optimal approach and method to establish fatigue accumulation of the wellhead, a monitoring data set is obtained on a well offshore West of Shetland. This paper presents an analysis of measured strain, motions and analytical predictions with the objective of better understanding the accuracy, limitations, or conservatism in each of the three methods defined above. Of the various parameters that affect the accuracy of the fatigue damage estimates, the paper identifies that the selection of analytical conductor-soil model is critical to narrowing the gap between fatigue life predictions from the different approaches. The work presented here presents the influence of alternative approaches to model conductor-soil interaction than the traditionally used API soil model. Overall, the paper presents the monitoring equipment and analytical methodology to advance the accuracy of wellhead fatigue damage measurements.


2018 ◽  
Vol 15 (6) ◽  
pp. 172988141881470
Author(s):  
Nezih Ergin Özkucur ◽  
H Levent Akın

Self-localization in autonomous robots is one of the fundamental issues in the development of intelligent robots, and processing of raw sensory information into useful features is an integral part of this problem. In a typical scenario, there are several choices for the feature extraction algorithm, and each has its weaknesses and strengths depending on the characteristics of the environment. In this work, we introduce a localization algorithm that is capable of capturing the quality of a feature type based on the local environment and makes soft selection of feature types throughout different regions. A batch expectation–maximization algorithm is developed for both discrete and Monte Carlo localization models, exploiting the probabilistic pose estimations of the robot without requiring ground truth poses and also considering different observation types as blackbox algorithms. We tested our method in simulations, data collected from an indoor environment with a custom robot platform and a public data set. The results are compared with the individual feature types as well as naive fusion strategy.


2018 ◽  
Vol 21 (7) ◽  
pp. 822-829 ◽  
Author(s):  
Mark Pennington ◽  
Richard Grieve ◽  
Jan Van der Meulen ◽  
Neil Hawkins
Keyword(s):  
Data Set ◽  

Risks ◽  
2021 ◽  
Vol 9 (11) ◽  
pp. 204
Author(s):  
Chamay Kruger ◽  
Willem Daniel Schutte ◽  
Tanja Verster

This paper proposes a methodology that utilises model performance as a metric to assess the representativeness of external or pooled data when it is used by banks in regulatory model development and calibration. There is currently no formal methodology to assess representativeness. The paper provides a review of existing regulatory literature on the requirements of assessing representativeness and emphasises that both qualitative and quantitative aspects need to be considered. We present a novel methodology and apply it to two case studies. We compared our methodology with the Multivariate Prediction Accuracy Index. The first case study investigates whether a pooled data source from Global Credit Data (GCD) is representative when considering the enrichment of internal data with pooled data in the development of a regulatory loss given default (LGD) model. The second case study differs from the first by illustrating which other countries in the pooled data set could be representative when enriching internal data during the development of a LGD model. Using these case studies as examples, our proposed methodology provides users with a generalised framework to identify subsets of the external data that are representative of their Country’s or bank’s data, making the results general and universally applicable.


2021 ◽  
Author(s):  
shenglan li ◽  
Zhuang Kang ◽  
jinyi Chen ◽  
Can Wang ◽  
Zehao Cai ◽  
...  

Abstract Background Medulloblastoma is a common intracranial tumor among children. In recent years, research on cancer genome has established four distinct subtypes of medulloblastoma: WNT, SHH, Group3, and Group4. Each subtype has its own transcriptional profile, methylation changes, and different clinical outcomes. Treatment and prognosis also vary depending on the subtype. Methods Based on the methylation data of medulloblastoma samples, methylCIBERSORT was used to evaluate the level of immune cell infiltration in medulloblastoma samples and identified 10 kinds of immune cells with different subtypes. Combined with the immune database, 293 Imm-DEGs were screened. Imm-DEGs were used to construct the co-expression network, and the key modules related to the level of differential immune cell infiltration were identified. Three immune hub genes (GAB1, ABL1, CXCR4) were identified according to the gene connectivity and the correlation with phenotype in the key modules, as well as the PPI network involved in the genes in the modules. Results The subtype marker was recognized according to the immune hub, and the subtype marker was verified in the external data set, the methylation level of immune hub gene among different subtypes was compared and analyzed, at the same time, tissue microarray was used for immunohistochemical verification, and a multi-factor regulatory network of hub gene was constructed. Conclusions Identifying subtype marker is helpful to accurately identify the subtypes of medulloblastoma patients, and can accurately evaluate the treatment and prognosis, so as to improve the overall survival of patients.


Sign in / Sign up

Export Citation Format

Share Document