scholarly journals Fuzzy Classification Rule Learning by Decision Tree Induction

Author(s):  
Keon-Myung Lee ◽  
Hak-Joon Kim
Author(s):  
KRZYSZTOF TRAWIŃSKI ◽  
OSCAR CORDÓN ◽  
ARNAUD QUIRIN

In this work, we conduct a study considering a fuzzy rule-based multiclassification system design framework based on Fuzzy Unordered Rule Induction Algorithm (FURIA). This advanced method serves as the fuzzy classification rule learning algorithm to derive the component classifiers considering bagging and feature selection. We develop an exhaustive study on the potential of bagging and feature selection to design a final FURIA-based fuzzy multiclassifier dealing with high dimensional data. Several parameter settings for the global approach are tested when applied to twenty one popular UCI datasets. The results obtained show that FURIA-based fuzzy multiclassifiers outperform the single FURIA classifier and are competitive with C4.5 multiclassifiers and random forests.


One of the important data mining functionality is classification. Presently, different methods exist for implementing classification. Rule-based classification using decision tree induction method is a conventional and simple method for identifying an unknown class of a given object. This method has a set of demerits and to remove these demerits, we depend on a soft computing tool which is known as soft set theory. One branch of soft set theory is called - multi soft theory- and it has a wide range of applications in the area of classification. We made a certain alteration in the rule-based classification using decision tree induction method by applying multi soft set theory. These changes will simplify the difficulties of the rule-based classification using decision tree induction method. The first two sections of this research work discuss introduction and preliminaries. In the remaining sections, the authors describe the multi soft set theory and its applications in rule base classification. Lastly, the paper finishes with a new algorithm, which the research scholars implemented as software using python programming. The suggested work experts can use in data mining industry. It has massive use in the fields of business, agriculture, health, education and many more.


Author(s):  
Ferdinand Bollwein ◽  
Stephan Westphal

AbstractUnivariate decision tree induction methods for multiclass classification problems such as CART, C4.5 and ID3 continue to be very popular in the context of machine learning due to their major benefit of being easy to interpret. However, as these trees only consider a single attribute per node, they often get quite large which lowers their explanatory value. Oblique decision tree building algorithms, which divide the feature space by multidimensional hyperplanes, often produce much smaller trees but the individual splits are hard to interpret. Moreover, the effort of finding optimal oblique splits is very high such that heuristics have to be applied to determine local optimal solutions. In this work, we introduce an effective branch and bound procedure to determine global optimal bivariate oblique splits for concave impurity measures. Decision trees based on these bivariate oblique splits remain fairly interpretable due to the restriction to two attributes per split. The resulting trees are significantly smaller and more accurate than their univariate counterparts due to their ability of adapting better to the underlying data and capturing interactions of attribute pairs. Moreover, our evaluation shows that our algorithm even outperforms algorithms based on heuristically obtained multivariate oblique splits despite the fact that we are focusing on two attributes only.


2021 ◽  
Vol 54 (1) ◽  
pp. 1-38
Author(s):  
Víctor Adrián Sosa Hernández ◽  
Raúl Monroy ◽  
Miguel Angel Medina-Pérez ◽  
Octavio Loyola-González ◽  
Francisco Herrera

Experts from different domains have resorted to machine learning techniques to produce explainable models that support decision-making. Among existing techniques, decision trees have been useful in many application domains for classification. Decision trees can make decisions in a language that is closer to that of the experts. Many researchers have attempted to create better decision tree models by improving the components of the induction algorithm. One of the main components that have been studied and improved is the evaluation measure for candidate splits. In this article, we introduce a tutorial that explains decision tree induction. Then, we present an experimental framework to assess the performance of 21 evaluation measures that produce different C4.5 variants considering 110 databases, two performance measures, and 10× 10-fold cross-validation. Furthermore, we compare and rank the evaluation measures by using a Bayesian statistical analysis. From our experimental results, we present the first two performance rankings in the literature of C4.5 variants. Moreover, we organize the evaluation measures into two groups according to their performance. Finally, we introduce meta-models that automatically determine the group of evaluation measures to produce a C4.5 variant for a new database and some further opportunities for decision tree models.


Author(s):  
Rodrigo C. Barros ◽  
Ricardo Cerri ◽  
Pablo A. Jaskowiak ◽  
Andre C. P. L. F. de Carvalho

Sign in / Sign up

Export Citation Format

Share Document