Correlation Analysis in Classifiers

Author(s):  
Vincent Lemaire ◽  
Carine Hue ◽  
Olivier Bernier

This chapter presents a new method to analyze the link between the probabilities produced by a classification model and the variation of its input values. The goal is to increase the predictive probability of a given class by exploring the possible values of the input variables taken independently. The proposed method is presented in a general framework, and then detailed for naive Bayesian classifiers. We also demonstrate the importance of “lever variables”, variables which can conceivably be acted upon to obtain specific results as represented by class probabilities, and consequently can be the target of specific policies. The application of the proposed method to several data sets shows that such an approach can lead to useful indicators.

Entropy ◽  
2018 ◽  
Vol 20 (11) ◽  
pp. 857 ◽  
Author(s):  
Khalil El Hindi ◽  
Hussien AlSalman ◽  
Safwan Qasem ◽  
Saad Al Ahmadi

Text classification is one domain in which the naive Bayesian (NB) learning algorithm performs remarkably well. However, making further improvement in performance using ensemble-building techniques proved to be a challenge because NB is a stable algorithm. This work shows that, while an ensemble of NB classifiers achieves little or no improvement in terms of classification accuracy, an ensemble of fine-tuned NB classifiers can achieve a remarkable improvement in accuracy. We propose a fine-tuning algorithm for text classification that is both more accurate and less stable than the NB algorithm and the fine-tuning NB (FTNB) algorithm. This improvement makes it more suitable than the FTNB algorithm for building ensembles of classifiers using bagging. Our empirical experiments, using 16-benchmark text-classification data sets, show significant improvement for most data sets.


Author(s):  
CHANG-HWAN LEE

In spite of its simplicity, naive Bayesian learning has been widely used in many data mining applications. However, the unrealistic assumption that all features are equally important negatively impacts the performance of naive Bayesian learning. In this paper, we propose a new method that uses a Kullback–Leibler measure to calculate the weights of the features analyzed in naive Bayesian learning. Its performance is compared to that of other state-of-the-art methods over a number of datasets.


2010 ◽  
Vol 105 (4) ◽  
pp. 435-466 ◽  
Author(s):  
Tayeb Kenaza ◽  
Karim Tabia ◽  
Salem Benferhat

2009 ◽  
Vol 2 (1) ◽  
pp. 1174-1185 ◽  
Author(s):  
Barzan Mozafari ◽  
Carlo Zaniolo

Sign in / Sign up

Export Citation Format

Share Document