From Local to Global Patterns: Evaluation Issues in Rule Learning Algorithms

Author(s):  
Johannes Fürnkranz
Author(s):  
Jamolbek Mattiev ◽  
Branko Kavsek

Huge amounts of data are being collected and analyzed nowadays. By using the popular rule-learning algorithms, the number of rule discovered on those ?big? datasets can easily exceed thousands. To produce compact, understandable and accurate classifiers, such rules have to be grouped and pruned, so that only a reasonable number of them are presented to the end user for inspection and further analysis. In this paper, we propose new methods that are able to reduce the number of class association rules produced by ?classical? class association rule classifiers, while maintaining an accurate classification model that is comparable to the ones generated by state-of-the-art classification algorithms. More precisely, we propose new associative classifiers, called DC, DDC and CDC, that use distance-based agglomerative hierarchical clustering as a post-processing step to reduce the number of its rules, and in the rule-selection step, we use different strategies (based on database coverage and cluster center) for each algorithm. Experimental results performed on selected datasets from the UCI ML repository show that our classifiers are able to learn classifiers containing significantly fewer rules than state-of-the-art rule learning algorithms on datasets with a larger number of examples. On the other hand, the classification accuracy of the proposed classifiers is not significantly different from state-of-the-art rule-learners on most of the datasets.


Author(s):  
Lianlong Wu

Declarative rules such as Prolog and Datalog are common formalisms to express expert knowledge and are used in a number of systems. Since developing such rules is time-consuming and requires scarce expert knowledge, it is essential to develop algorithms for learning such rules. We address the problem of learning existential rules, a richer class of rules which found applications in many use-cases such as Semantic Web and Web Data Extraction. In particular, we concentrate on developing evolutionary learning algorithms for rule learning.


Author(s):  
A. GONZÁLEZ ◽  
R. PÉREZ

A very important problem associated with the use of learning algorithms consists of fixing the correct assignment of the initial domains for the predictive variables. In the fuzzy case, this problem is equivalent of define the fuzzy labels for each variable. In this work, we propose the inclusion in a learning algorithm, called SLAVE, of a particular kind of linguistic hedges as a way to modify the intial semantic of the labels. These linguistic hedges allow us both to learn and to tune fuzzy rules.


1998 ◽  
Vol 8 ◽  
pp. 129-164 ◽  
Author(s):  
J. Fuernkranz

In this paper we re-investigate windowing for rule learning algorithms. We show that, contrary to previous results for decision tree learning, windowing can in fact achieve significant run-time gains in noise-free domains and explain the different behavior of rule learning algorithms by the fact that they learn each rule independently. The main contribution of this paper is integrative windowing, a new type of algorithm that further exploits this property by integrating good rules into the final theory right after they have been discovered. Thus it avoids re-learning these rules in subsequent iterations of the windowing process. Experimental evidence in a variety of noise-free domains shows that integrative windowing can in fact achieve substantial run-time gains. Furthermore, we discuss the problem of noise in windowing and present an algorithm that is able to achieve run-time gains in a set of experiments in a simple domain with artificial noise.


Sign in / Sign up

Export Citation Format

Share Document