Rule extraction from electroencephalogram signals using support vector machine

Author(s):  
Anuchin Chatchinarat ◽  
Kok Wai Wong ◽  
Chun Che Fung
2013 ◽  
Vol 572 ◽  
pp. 300-303
Author(s):  
Jian Guo Wang ◽  
Bin Yang ◽  
Wen Xing Zhang ◽  
Bo Qin

A new rule extraction algorithm based on convex hull for strip hot-dip galvanizing process monitoring is proposed in this paper. It overcomes the black-box problem of support vector machine. The zinc coating weight is used as the investigated subject. The sample datasets are trained by support vector machine rule extraction method, and the quantitative relationship can be obtained in the form of knowledge rules among input variables (such as the parameters of raw materials and control parameters of production) and output ones (the quality parameters), with which the production control parameters can be set and updated easily.


2008 ◽  
pp. 1269-1279
Author(s):  
Xiuju Fu ◽  
Lipo Wang ◽  
GihGuang Hung ◽  
Liping Goh

Classification decisions from linguistic rules are more desirable compared to complex mathematical formulas from support vector machine (SVM) classifiers due to the explicit explanation capability of linguistic rules. Linguistic rule extraction has been attracting much attention in explaining knowledge hidden in data. In this chapter, we show that the decisions from an SVM classifier can be decoded into linguistic rules based on the information provided by support vectors and decision function. Given a support vector of a certain class, cross points between each line, which is extended from the support vector along each axis, and an SVM decision hyper-curve are searched first. A hyper-rectangular rule is derived from these cross points. The hyper-rectangle is tuned by a tuning phase in order to exclude those out-class data points. Finally, redundant rules are merged to produce a compact rule set. Simultaneously, important attributes could be highlighted in the extracted rules. Rule extraction results from our proposed method could follow SVM classifier decisions very well. We compare the rule extraction results from SVM with RBF kernel function and linear kernel function. Experiment results show that rules extracted from SVM with RBF nonlinear kernel function are with better accuracy than rules extracted from SVM with linear kernel function. Comparisons between our method and other rule extraction methods are also carried out on several benchmark data sets. Higher rule accuracy is obtained in our method with fewer number of premises in each rule.


Author(s):  
Xiuju Fu ◽  
Lipo Wang ◽  
GihGuang Hung ◽  
Liping Goh

Classification decisions from linguistic rules are more desirable compared to complex mathematical formulas from support vector machine (SVM) classifiers due to the explicit explanation capability of linguistic rules. Linguistic rule extraction has been attracting much attention in explaining knowledge hidden in data. In this chapter, we show that the decisions from an SVM classifier can be decoded into linguistic rules based on the information provided by support vectors and decision function. Given a support vector of a certain class, cross points between each line, which is extended from the support vector along each axis, and an SVM decision hyper-curve are searched first. A hyper-rectangular rule is derived from these cross points. The hyper-rectangle is tuned by a tuning phase in order to exclude those out-class data points. Finally, redundant rules are merged to produce a compact rule set. Simultaneously, important attributes could be highlighted in the extracted rules. Rule extraction results from our proposed method could follow SVM classifier decisions very well. We compare the rule extraction results from SVM with RBF kernel function and linear kernel function. Experiment results show that rules extracted from SVM with RBF nonlinear kernel function are with better accuracy than rules extracted from SVM with linear kernel function. Comparisons between our method and other rule extraction methods are also carried out on several benchmark data sets. Higher rule accuracy is obtained in our method with fewer number of premises in each rule.


Sign in / Sign up

Export Citation Format

Share Document