Reseach on Feature Selection Algorithm Based on the Margin of Support Vector Machine

2013 ◽  
Vol 333-335 ◽  
pp. 1430-1434
Author(s):  
Lin Fang Hu ◽  
Lei Qiao ◽  
Min De Huang

A feature selection algorithm based on the optimal hyperplane of SVM is raised. Using the algorithm, the contribution to the classification of each feature in the candidate feature set is test, and then the feature subset with best classification ability will be selected. The algorithm is used in the recognition process of storm monomers in weather forecast, and experimental data show that the classification ability of the features can be effectively evaluated; the optimal feature subset is selected to enhance the working performance of the classifier.

Author(s):  
Hui Wang ◽  
Li Li Guo ◽  
Yun Lin

Automatic modulation recognition is very important for the receiver design in the broadband multimedia communication system, and the reasonable signal feature extraction and selection algorithm is the key technology of Digital multimedia signal recognition. In this paper, the information entropy is used to extract the single feature, which are power spectrum entropy, wavelet energy spectrum entropy, singular spectrum entropy and Renyi entropy. And then, the feature selection algorithm of distance measurement and Sequential Feature Selection(SFS) are presented to select the optimal feature subset. Finally, the BP neural network is used to classify the signal modulation. The simulation result shows that the four-different information entropy can be used to classify different signal modulation, and the feature selection algorithm is successfully used to choose the optimal feature subset and get the best performance.


2013 ◽  
Vol 347-350 ◽  
pp. 2712-2716
Author(s):  
Lin Tao Lü ◽  
Peng Li ◽  
Yu Xiang Yang ◽  
Fang Tan

According to the features of Palm bio-impedance spectroscopy (BIS) data, this paper suggests a kind of effective feature model of palm BIS data elliptical model. The model combines immune clone algorithm and least squares method, establishes a palm BIS feature selection algorithm, and uses the algorithm to obtain the optimal feature subset that can completely represent the palm BIS data, and then use several classification algorithms for classification and comparison. The experimental results show that accuracy of the feature subset obtained through the algorithm in SVM classification algorithm test can reach 93.2, thereby verifying the algorithm is a valid and reliable palm BIS feature selection algorithm.


2020 ◽  
Vol 30 (11) ◽  
pp. 2050017 ◽  
Author(s):  
Jian Lian ◽  
Yunfeng Shi ◽  
Yan Zhang ◽  
Weikuan Jia ◽  
Xiaojun Fan ◽  
...  

Feature selection plays a vital role in the detection and discrimination of epileptic seizures in electroencephalogram (EEG) signals. The state-of-the-art EEG classification techniques commonly entail the extraction of the multiple features that would be fed into classifiers. For some techniques, the feature selection strategies have been used to reduce the dimensionality of the entire feature space. However, most of these approaches focus on the performance of classifiers while neglecting the association between the feature and the EEG activity itself. To enhance the inner relationship between the feature subset and the epileptic EEG task with a promising classification accuracy, we propose a machine learning-based pipeline using a novel feature selection algorithm built upon a knockoff filter. First, a number of temporal, spectral, and spatial features are extracted from the raw EEG signals. Second, the proposed feature selection algorithm is exploited to obtain the optimal subgroup of features. Afterwards, three classifiers including [Formula: see text]-nearest neighbor (KNN), random forest (RF) and support vector machine (SVM) are used. The experimental results on the Bonn dataset demonstrate that the proposed approach outperforms the state-of-the-art techniques, with accuracy as high as 99.93% for normal and interictal EEG discrimination and 98.95% for interictal and ictal EEG classification. Meanwhile, it has achieved satisfactory sensitivity (95.67% in average), specificity (98.83% in average), and accuracy (98.89% in average) over the Freiburg dataset.


2013 ◽  
Vol 774-776 ◽  
pp. 1532-1537
Author(s):  
Jing Wei Yang ◽  
Si Le Wang ◽  
Ying Yi Chen ◽  
Su Kui Lu ◽  
Wen Zhu Yang

This paper presents a genetic-based feature selection algorithm for object recognition. Firstly, the proposed algorithm encodes a solution with a binary chromosome. Secondly, the initial population was generated randomly. Thirdly, a crossover operator and a mutation operator are employed to operate on these chromosomes to generate more competency chromosomes. The probability of the crossover and mutation are adjusted dynamically according to the generation number and the fitness value. The proposed algorithm is tested using the features extracted from cotton foreign fiber objects. The results indicate that the proposed algorithm can obtain the optimal feature subset, and can reduce the classification time while keeping the classification accuracy constant.


2015 ◽  
Vol 16 (2) ◽  
pp. 363
Author(s):  
LinTao LV ◽  
QinQin Yuan ◽  
YuXiang Yang

A kind of effective feature model of palm BIS data is presented according to the features of Palm bio-impedance spectroscopy (BIS) data. Based on immune clone algorithm and least squares method, an improved palm BIS feature selection algorithm is established,which can be applied to obtain the optimal feature subset that can be completely represented the palm BIS data, Finally, the algorithm is compared with other algorithms. The experimental results show that the accuracy of the feature subset obtained by the algorithm has reached 93.2 in SVM classification algorithm test.Therefore, the algorithm in this article is valid and reliable , which is of high theoretical and practical value.


2015 ◽  
Vol 2015 ◽  
pp. 1-9 ◽  
Author(s):  
Senthilkumar Devaraj ◽  
S. Paulraj

Multidimensional medical data classification has recently received increased attention by researchers working on machine learning and data mining. In multidimensional dataset (MDD) each instance is associated with multiple class values. Due to its complex nature, feature selection and classifier built from the MDD are typically more expensive or time-consuming. Therefore, we need a robust feature selection technique for selecting the optimum single subset of the features of the MDD for further analysis or to design a classifier. In this paper, an efficient feature selection algorithm is proposed for the classification of MDD. The proposed multidimensional feature subset selection (MFSS) algorithm yields a unique feature subset for further analysis or to build a classifier and there is a computational advantage on MDD compared with the existing feature selection algorithms. The proposed work is applied to benchmark multidimensional datasets. The number of features was reduced to 3% minimum and 30% maximum by using the proposed MFSS. In conclusion, the study results show that MFSS is an efficient feature selection algorithm without affecting the classification accuracy even for the reduced number of features. Also the proposed MFSS algorithm is suitable for both problem transformation and algorithm adaptation and it has great potentials in those applications generating multidimensional datasets.


Author(s):  
Smita Chormunge ◽  
Sudarson Jena

<p>Feature selection approach solves the dimensionality problem by removing irrelevant and redundant features. Existing Feature selection algorithms take more time to obtain feature subset for high dimensional data. This paper proposes a feature selection algorithm based on Information gain measures for high dimensional data termed as IFSA (Information gain based Feature Selection Algorithm) to produce optimal feature subset in efficient time and improve the computational performance of learning algorithms. IFSA algorithm works in two folds: First apply filter on dataset. Second produce the small feature subset by using information gain measure. Extensive experiments are carried out to compare proposed algorithm and other methods with respect to two different classifiers (Naive bayes and IBK) on microarray and text data sets. The results demonstrate that IFSA not only produces the most select feature subset in efficient time but also improves the classifier performance.</p>


Author(s):  
Smita Chormunge ◽  
Sudarson Jena

<p>Feature selection approach solves the dimensionality problem by removing irrelevant and redundant features. Existing Feature selection algorithms take more time to obtain feature subset for high dimensional data. This paper proposes a feature selection algorithm based on Information gain measures for high dimensional data termed as IFSA (Information gain based Feature Selection Algorithm) to produce optimal feature subset in efficient time and improve the computational performance of learning algorithms. IFSA algorithm works in two folds: First apply filter on dataset. Second produce the small feature subset by using information gain measure. Extensive experiments are carried out to compare proposed algorithm and other methods with respect to two different classifiers (Naive bayes and IBK) on microarray and text data sets. The results demonstrate that IFSA not only produces the most select feature subset in efficient time but also improves the classifier performance.</p>


Sign in / Sign up

Export Citation Format

Share Document