scholarly journals Multi-Seam Coal Deposit Modeling via Principal Component Analysis & GIS

Minerals ◽  
2019 ◽  
Vol 9 (9) ◽  
pp. 532
Author(s):  
Georgios Louloudis ◽  
Christos Roumpos ◽  
Konstantinos Theofilogiannakos ◽  
Nikolaos Stathopoulos

Spatial modeling and evaluation is a critical step for planning the exploitation of mineral deposits. In this work, a methodology for the investigation of a multi-seam coal deposit spatial variability is proposed. The study area includes the Klidi (Florina, Greece) multi-seam lignite deposit which is suitable for surface mining. The analysis is based on the original data of 76 exploratory drill-holes in an area of 10 km2, in conjunction with the geological and geomorphological data of the deposit. The analytical methods include drill-hole data analysis and evaluation based on an appropriate algorithm, principal component analysis and geographic information techniques. The results proved to be very satisfactory for the explanation of the maximum variance of the initial data values as well as the identification of the deposit structure and the optimum planning of mine development. The proposed analysis can be also helpful for minimizing cost and optimizing efficiency of surface mining operations. Furthermore, the provided methods could be applied in other areas of geosciences, indicating the theoretical value as well as the important practical implications of the analysis.

2020 ◽  
Vol 23 (11) ◽  
pp. 2414-2430
Author(s):  
Khaoula Ghoulem ◽  
Tarek Kormi ◽  
Nizar Bel Hadj Ali

In the general framework of data-driven structural health monitoring, principal component analysis has been applied successfully in continuous monitoring of complex civil infrastructures. In the case of linear or polynomial relationship between monitored variables, principal component analysis allows generation of structured residuals from measurement outputs without a priori structural model. The principal component analysis has been widely used for system monitoring based on its ability to handle high-dimensional, noisy, and highly correlated data by projecting the data onto a lower dimensional subspace that contains most of the variance of the original data. However, for nonlinear systems, it could be easily demonstrated that linear principal component analysis is unable to disclose nonlinear relationships between variables. This has naturally motivated various developments of nonlinear principal component analysis to tackle damage diagnosis of complex structural systems, especially those characterized by a nonlinear behavior. In this article, a data-driven technique for damage detection in nonlinear structural systems is presented. The proposed method is based on kernel principal component analysis. Two case studies involving nonlinear cable structures are presented to show the effectiveness of the proposed methodology. The validity of the kernel principal component analysis–based monitoring technique is shown in terms of the ability to damage detection. Robustness to environmental effects and disturbances are also studied.


2019 ◽  
Vol 12 (1) ◽  
pp. 23
Author(s):  
Alissar Nasser

We study in this paper the performance of Hospitals in Lebanon. Using the nonparametric method Data Envelopment Analysis (DEA), we are able to measures relative efficiency of Hospitals in Lebanon. DEA is a technique that uses linear programming and it measures the relative efficiency of similar type of organizations termed as Decision Making Units (DMUs). In this study, due to the lack of individual data on hospital level, each DMU refers to a qada in Lebanon where the used data represent the aggregation of input and outputs of different hospitals within the qada. In DEA, the inclusion of more number of inputs and /or outputs results in getting a more number of efficient units. Therefore, selecting the appropriate inputs and outputs is a major factor of DEA results. Therefore, we use here the Principal Component Analysis (PCA) in order to reduce the data structure into certain principal components which are essential for identifying efficient DMUs. It is important to note that we have used the basic BCC-input model for the entire analysis. We considered 24 DMUs for the study, using DEA on original data; we got 17 DMUs out of 24 DMUs as efficient. Then we considered 1 PC for inputs and 1 PC for output with almost 80 percent variances, resulting in 3 DMUs as efficient and 21 as inefficient. Using 1 PC for input and 2 PCs for output with 90 percent variance for both input and output, we got 9 DMUs as efficient and 15 DMUs as inefficient. Finally, we have attempted to identify the efficient units with 2 PCs and for 2 PCs for input and outputs with variance more than 95 percent, resulting in 10 efficient DMUs and 14 inefficient DMUs. In Principal Component analysis, if the variance lies between 80 percent to-90 percent it is judged as a meaningful one. It is concluded that Principal Component Analysis plays an important role in the reduction of input output variables and helps in identifying the efficient DMUs and improves the discriminating power of DEA.


2011 ◽  
Vol 15 (1) ◽  
pp. 178
Author(s):  
Altien J Rindengan ◽  
Deiby Tineke Salaki

PENGELOMPOKKAN DATA WAJAH MENGGUNAKAN METODE AGGLOMERATIVE CLUSTERING DENGAN ANALISIS KOMPONEN UTAMA Altien J. Rindengan1) dan Deiby Tineke Salaki1) 1)Program Studi Matematika FMIPA Universitas Sam Ratulangi Manado 95115 ABSTRAK Pada penelitian ini dilakukan analisis pengelompokkan data wajah dengan analisis komponen utama untuk mengambil beberapa akar ciri yang cukup mewakili data tersebut dan pengelompokkannya menggunakan metode agglomerative clustering. Dengan menggunakan program Matlab, data wajah yang terdiri dari 6 orang dengan 10 image dapat dikelompokkan sesuai data aslinya.  Pengelompokkannya cukup menggunakan 3 akar ciri pada selang 68 %. Kata kunci: agglomerative clustering, analisis komponen utama, data wajah  FACE DATA CLUSTERING USING AGGLOMERATIVE CLUSTERING METHODS WITH PRINCIPAL COMPONENT ANALYSIS ABSTRACT In this research, face data is grouped using principal component analysis by getting some of its eigenvalues which are representative enough to describe the data and then by using agglomerative clustering the data is clustered.  By running the Matlab program, face data which is consist of 6 people with 10 images can be clustered to fit the original data.  The clustering is enough using 3 eigenvalues with 68 % of interval. Keywords: agglomerative clustering, principal component analysis, face data


2020 ◽  
Vol 23 ◽  
pp. 41-44
Author(s):  
Oļegs Užga-Rebrovs ◽  
Gaļina Kuļešova

Any data in an implicit form contain information of interest to the researcher. The purpose of data analysis is to extract this information. The original data may contain redundant elements and noise, distorting these data to one degree or another. Therefore, it seems necessary to subject the data to preliminary processing. Reducing the dimension of the initial data makes it possible to remove interfering factors and present the data in a form suitable for further analysis. The paper considers an approach to reducing the dimensionality of the original data based on principal component analysis.


2020 ◽  
Vol 2020 ◽  
pp. 1-12
Author(s):  
Pei Heng Li ◽  
Taeho Lee ◽  
Hee Yong Youn

Various dimensionality reduction (DR) schemes have been developed for projecting high-dimensional data into low-dimensional representation. The existing schemes usually preserve either only the global structure or local structure of the original data, but not both. To resolve this issue, a scheme called sparse locality for principal component analysis (SLPCA) is proposed. In order to effectively consider the trade-off between the complexity and efficiency, a robust L2,p-norm-based principal component analysis (R2P-PCA) is introduced for global DR, while sparse representation-based locality preserving projection (SR-LPP) is used for local DR. Sparse representation is also employed to construct the weighted matrix of the samples. Being parameter-free, this allows the construction of an intrinsic graph more robust against the noise. In addition, simultaneous learning of projection matrix and sparse similarity matrix is possible. Experimental results demonstrate that the proposed scheme consistently outperforms the existing schemes in terms of clustering accuracy and data reconstruction error.


Author(s):  
Xin Li

In this paper, we present approaches to perform principal component analysis (PCA) clustering for distributed heterogeneous genomic datasets with privacy protection. The approaches allow data providers to collaborate together to identify gene profiles from a global viewpoint, and at the same time, protect the sensitive genomic data from possible privacy leaks. We then further develop a framework for privacy preserving PCA-based gene clustering, which includes two types of participants: data providers and a trusted central site (TCS). Two different methodologies are employed: Collective PCA (C-PCA) and Repeating PCA (R-PCA). The C-PCA requires local sites to transmit a sample of original data to the TCS and can be applied to any heterogeneous datasets. The R-PCA approach requires all local sites have the same or similar number of columns, but releases no original data. Experiments on five independent genomic datasets show that both C-PCA and R-PCA approaches maintain very good accuracy compared with the centralized scenario.


2018 ◽  
Author(s):  
Travis W. Tokarek ◽  
Charles A. Odame-Ankrah ◽  
Jennifer A. Huo ◽  
Robert McLaren ◽  
Alex K. Y. Lee ◽  
...  

Abstract. In this paper, measurements of air pollutants made at a ground site near Fort McKay in the Athabasca oil sands region as part of a multi-platform campaign in the summer of 2013 are presented. The observations included measurements of selected volatile organic compounds (VOCs) by a gas chromatograph &ndash ion trap mass spectrometer (GC-ITMS). This instrument observed a large, analytically unresolved hydrocarbon peak (with retention index between 1100 and 1700) associated with intermediate volatility organic compounds (IVOCs). However, the activities or processes that contribute to the release of these IVOCs in the oil sands region remain unclear. Principal component analysis (PCA) with Varimax rotation was applied to elucidate major source types impacting the sampling site in the summer of 2013. The analysis included 28 variables, including concentrations of total odd nitrogen (NOy), carbon dioxide (CO2), methane (CH4), ammonia (NH3), carbon monoxide (CO), sulfur dioxide (SO2), total reduced sulfur compounds (TRS), speciated monoterpenes (including α- and β-pinene and limonene), particle volume calculated from measured size distributions of particles less than 10 µm and 1 µm in diameter (PM10-1 and PM1), particle-surface bound polycyclic aromatic hydrocarbons (pPAH), and aerosol mass spectrometer composition measurements, including refractory black carbon (rBC) and organic aerosol components. The PCA was complemented by bivariate polar plots showing the joint wind speed and direction dependence of air pollutant concentrations to illustrate the spatial distribution of sources in the area. Using the 95 % cumulative percentage of variance criterion, ten components were identified and categorized by source type. These included emissions by wet tailings ponds, vegetation, open pit mining operations, upgrader facilities, and surface dust. Three components correlated with IVOCs, with the largest associated with surface mining and is likely caused by the unearthing and processing of raw bitumen.


2002 ◽  
Vol 30 (4) ◽  
pp. 239-247
Author(s):  
S. M. Shamsuddin ◽  
M. Darus ◽  
M. N. Sulaiman

Data reduction is a process of feature extraction that transforms the data space into a feature space of much lower dimension compared to the original data space, yet it retains most of the intrinsic information content of the data. This can be done by using a number of methods, such as principal component analysis (PCA), factor analysis, and feature clustering. Principal components are extracted from a collection of multivariate cases as a way of accounting for as much of the variation in that collection as possible by means of as few variables as possible. On the other hand, backpropagation network has been used extensively in classification problems such as XOR problems, share prices prediction, and pattern recognition. This paper proposes an improved error signal of backpropagation network for classification of the reduction invariants using principal component analysis, for extracting the bulk of the useful information present in moment invariants of handwritten digits, leaving the redundant information behind. Higher order centralised scale- invariants are used to extract features of handwritten digits before PCA, and the reduction invariants are sent to the improved backpropagation model for classification purposes.


Author(s):  
Xin Li

In this paper, we present approaches to perform principal component analysis (PCA) clustering for distributed heterogeneous genomic datasets with privacy protection. The approaches allow data providers to collaborate together to identify gene profiles from a global viewpoint, and at the same time, protect the sensitive genomic data from possible privacy leaks. We then further develop a framework for privacy preserving PCA-based gene clustering, which includes two types of participants: data providers and a trusted central site (TCS). Two different methodologies are employed: Collective PCA (C-PCA) and Repeating PCA (R-PCA). The C-PCA requires local sites to transmit a sample of original data to the TCS and can be applied to any heterogeneous datasets. The R-PCA approach requires all local sites have the same or similar number of columns, but releases no original data. Experiments on five independent genomic datasets show that both C-PCA and R-PCA approaches maintain very good accuracy compared with the centralized scenario.


2007 ◽  
Vol 7 (3) ◽  
pp. 875-886 ◽  
Author(s):  
T. W. Chan ◽  
M. Mozurkewich

Abstract. Principal component analysis provides a fast and robust method to reduce the data dimensionality of an aerosol size distribution data set. Here we describe a methodology for applying principal component analysis to aerosol size distribution measurements. We illustrate the method by applying it to data obtained during five field studies. Most variations in the sub-micrometer aerosol size distribution over periods of weeks can be described using 5 components. Using 6 to 8 components preserves virtually all the information in the original data. A key aspect of our approach is the introduction of a new method to weight the data; this preserves the orthogonality of the components while taking the measurement uncertainties into account. We also describe a new method for identifying the approximate number of aerosol components needed to represent the measurement quantitatively. Applying Varimax rotation to the resultant components decomposes a distribution into independent monomodal distributions. Normalizing the components provides physical meaning to the component scores. The method is relatively simple, computationally fast, and numerically robust. The resulting data simplification provides an efficient method of representing complex data sets and should greatly assist in the analysis of size distribution data.


Sign in / Sign up

Export Citation Format

Share Document