Nonlinear and additive principal component analysis for functional data

2021 ◽  
Vol 181 ◽  
pp. 104675
Author(s):  
Jun Song ◽  
Bing Li
Author(s):  
Peter Hall

This article discusses the methodology and theory of principal component analysis (PCA) for functional data. It first provides an overview of PCA in the context of finite-dimensional data and infinite-dimensional data, focusing on functional linear regression, before considering the applications of PCA for functional data analysis, principally in cases of dimension reduction. It then describes adaptive methods for prediction and weighted least squares in functional linear regression. It also examines the role of principal components in the assessment of density for functional data, showing how principal component functions are linked to the amount of probability mass contained in a small ball around a given, fixed function, and how this property can be used to define a simple, easily estimable density surrogate. The article concludes by explaining the use of PCA for estimating log-density.


2016 ◽  
Vol 63 (1) ◽  
pp. 81-97
Author(s):  
Mirosław Krzyśko ◽  
Agnieszka Majka ◽  
Waldemar Wołyński

The paper presents an estimation of life standard diversity for residents of Polish voivodships in 2003–2013. The principal component analysis was applied for multidimensional functional data and the dendrite method was used for cluster analysis. These methods made it possible to isolate relatively homogeneous groups of voivodships that had similar values of characteristics under consideration, for the whole period at issue.


Symmetry ◽  
2020 ◽  
Vol 12 (1) ◽  
pp. 182
Author(s):  
Fengmin Yu ◽  
Liming Liu ◽  
Nanxiang Yu ◽  
Lianghao Ji ◽  
Dong Qiu

Recently, with the popularization of intelligent terminals, research on intelligent big data has been paid more attention. Among these data, a kind of intelligent big data with functional characteristics, which is called functional data, has attracted attention. Functional data principal component analysis (FPCA), as an unsupervised machine learning method, plays a vital role in the analysis of functional data. FPCA is the primary step for functional data exploration, and the reliability of FPCA plays an important role in subsequent analysis. However, classical L2-norm functional data principal component analysis (L2-norm FPCA) is sensitive to outliers. Inspired by the multivariate data L1-norm principal component analysis methods, we propose an L1-norm functional data principal component analysis method (L1-norm FPCA). Because the proposed method utilizes L1-norm, the L1-norm FPCs are less sensitive to the outliers than L2-norm FPCs which are the characteristic functions of symmetric covariance operator. A corresponding algorithm for solving the L1-norm maximized optimization model is extended to functional data based on the idea of the multivariate data L1-norm principal component analysis method. Numerical experiments show that L1-norm FPCA proposed in this paper has a better robustness than L2-norm FPCA, and the reconstruction ability of the L1-norm principal component analysis to the original uncontaminated functional data is as good as that of the L2-norm principal component analysis.


Author(s):  
Zhenhua Lin ◽  
Hongtu Zhu

We consider the problem of performing dimension reduction on heteroscedastic functional data where the variance is in different scales over entire domain. The aim of this paper is to propose a novel multiscale functional principal component analysis (MFPCA) approach to address such heteroscedastic issue. The key ideas of MFPCA are to partition the whole domain into several subdomains according to the scale of variance, and then to conduct the usual functional principal component analysis (FPCA) on each individual subdomain. Both theoretically and numerically, we show that MFPCA can capture features on areas of low variance without estimating high-order principal components, leading to overall improvement of performance on dimension reduction for heteroscedastic functional data. In contrast, traditional FPCA prioritizes optimizing performance on the subdomain of larger data variance and requires a practically prohibitive number of components to characterize data in the region bearing relatively small variance.


Sign in / Sign up

Export Citation Format

Share Document