Application of Kernel Trick to Fuzzy c-Means with Regularization by K-L Information

Author(s):  
Hidetomo Ichihashi ◽  
◽  
Katsuhiro Honda

Support vector machines (SVM), kernel principal component analysis (KPCA), and kernel Fisher discriminant analysis (KFD), are examples of successful kernel-based learning methods. By the addition of a regularizer and the kernel trick to a fuzzy counterpart of Gaussian mixture models (GMM), this paper proposes a clustering algorithm in an extended high dimensional feature space. Unlike the global nonlinear approaches, GMM or its fuzzy counterpart is to model nonlinear structure with a collection, or mixture, of local linear sub-models of PCA. When the number of feature vectors and clusters are n and c respectively, this kernel approach can find up to c × n nonzero eigenvalues. A way to control the number of parameters in the mixture of probabilistic principal component analysis (PPCA) is adopted to reduce the number of parameters. The algorithm provides a partitioning with flexible shape of clusters in the original input data space.

Entropy ◽  
2018 ◽  
Vol 20 (9) ◽  
pp. 701 ◽  
Author(s):  
Beige Ye ◽  
Taorong Qiu ◽  
Xiaoming Bai ◽  
Ping Liu

In view of the nonlinear characteristics of electroencephalography (EEG) signals collected in the driving fatigue state recognition research and the issue that the recognition accuracy of the driving fatigue state recognition method based on EEG is still unsatisfactory, this paper proposes a driving fatigue recognition method based on sample entropy (SE) and kernel principal component analysis (KPCA), which combines the advantage of the high recognition accuracy of sample entropy and the advantages of KPCA in dimensionality reduction for nonlinear principal components and the strong non-linear processing capability. By using support vector machine (SVM) classifier, the proposed method (called SE_KPCA) is tested on the EEG data, and compared with those based on fuzzy entropy (FE), combination entropy (CE), three kinds of entropies including SE, FE and CE that merged with KPCA. Experiment results show that the method is effective.


2005 ◽  
Vol 2005 (2) ◽  
pp. 155-159 ◽  
Author(s):  
Zhenqiu Liu ◽  
Dechang Chen ◽  
Halima Bensmail

One important feature of the gene expression data is that the number of genesMfar exceeds the number of samplesN. Standard statistical methods do not work well whenN<M. Development of new methodologies or modification of existing methodologies is needed for the analysis of the microarray data. In this paper, we propose a novel analysis procedure for classifying the gene expression data. This procedure involves dimension reduction using kernel principal component analysis (KPCA) and classification with logistic regression (discrimination). KPCA is a generalization and nonlinear version of principal component analysis. The proposed algorithm was applied to five different gene expression datasets involving human tumor samples. Comparison with other popular classification methods such as support vector machines and neural networks shows that our algorithm is very promising in classifying gene expression data.


Sign in / Sign up

Export Citation Format

Share Document