scholarly journals Learning optimization for decision tree classification of non-categorical data with information gain impurity criterion

Author(s):  
K. I. Sofeikov ◽  
I. Yu. Tyukin ◽  
A. N. Gorban ◽  
E. M. Mirkes ◽  
D. V. Prokhorov ◽  
...  
2019 ◽  
Vol 5 ◽  
pp. 147-152
Author(s):  
Subik Shrestha ◽  
Laxman Paudel

There is a possibility in finding hidden patterns that might help find a relationship between the information provided by the Loan Applicants during the Loan Application process and the status of their loan repayment. This paper highlights on finding such patterns by building a Decision Tree with the help of the data provided during the loan application process. Eleven attribute information of Five Hundred sixty four loan applicants were collected from Garima Bikas Bank Ltd. A decision tree model with a depth of 6 has been built by calculating the entropy and information gain at each split and selecting the feature with the highest information gain.


2021 ◽  
pp. 1-10
Author(s):  
Chao Dong ◽  
Yan Guo

The wide application of artificial intelligence technology in various fields has accelerated the pace of people exploring the hidden information behind large amounts of data. People hope to use data mining methods to conduct effective research on higher education management, and decision tree classification algorithm as a data analysis method in data mining technology, high-precision classification accuracy, intuitive decision results, and high generalization ability make it become a more ideal method of higher education management. Aiming at the sensitivity of data processing and decision tree classification to noisy data, this paper proposes corresponding improvements, and proposes a variable precision rough set attribute selection standard based on scale function, which considers both the weighted approximation accuracy and attribute value of the attribute. The number improves the anti-interference ability of noise data, reduces the bias in attribute selection, and improves the classification accuracy. At the same time, the suppression factor threshold, support and confidence are introduced in the tree pre-pruning process, which simplifies the tree structure. The comparative experiments on standard data sets show that the improved algorithm proposed in this paper is better than other decision tree algorithms and can effectively realize the differentiated classification of higher education management.


2005 ◽  
Vol 82 (12) ◽  
pp. 1038-1046 ◽  
Author(s):  
MICHAEL D. TWA ◽  
SRINIVASAN PARTHASARATHY ◽  
CYNTHIA ROBERTS ◽  
ASHRAF M. MAHMOUD ◽  
THOMAS W. RAASCH ◽  
...  

2019 ◽  
Vol 46 (3) ◽  
pp. 325-339
Author(s):  
Muhammad Shaheen ◽  
Tanveer Zafar ◽  
Sajid Ali Khan

Selection of an attribute for placement of the decision tree at an appropriate position (e.g. root of the tree) is an important decision. Many attribute selection measures such as Information Gain, Gini Index and Entropy have been developed for this purpose. The suitability of an attribute generally depends on the diversity of its values, relevance and dependency. Different attribute selection measures have different criteria for measuring the suitability of an attribute. Diversity Index is a classical statistical measure for determining the diversity of values, and according to our knowledge, it has never been used as an attribute selection method. In this article, we propose a novel attribute selection method for decision tree classification. In the proposed scheme, the average of Information Gain, Gini Index and Diversity Index are taken into account for assigning a weight to the attributes. The attribute with the highest average value is selected for the classification. We have empirically tested our proposed algorithm for classification of different data sets of scientific journals and conferences. We have developed a web-based application named JC-Rank that makes use of our proposed algorithm. We have also compared the results of our proposed technique with some existing decision tree classification algorithms.


Author(s):  
Dympna O’Sullivan ◽  
William Elazmeh ◽  
Szymon Wilk ◽  
Ken Farion ◽  
Stan Matwin ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document