scholarly journals On cognitive preferences and the plausibility of rule-based models

2019 ◽  
Vol 109 (4) ◽  
pp. 853-898 ◽  
Author(s):  
Johannes Fürnkranz ◽  
Tomáš Kliegr ◽  
Heiko Paulheim

AbstractIt is conventional wisdom in machine learning and data mining that logical models such as rule sets are more interpretable than other models, and that among such rule-based models, simpler models are more interpretable than more complex ones. In this position paper, we question this latter assumption by focusing on one particular aspect of interpretability, namely the plausibility of models. Roughly speaking, we equate the plausibility of a model with the likeliness that a user accepts it as an explanation for a prediction. In particular, we argue that—all other things being equal—longer explanations may be more convincing than shorter ones, and that the predominant bias for shorter models, which is typically necessary for learning powerful discriminative models, may not be suitable when it comes to user acceptance of the learned models. To that end, we first recapitulate evidence for and against this postulate, and then report the results of an evaluation in a crowdsourcing study based on about 3000 judgments. The results do not reveal a strong preference for simple rules, whereas we can observe a weak preference for longer rules in some domains. We then relate these results to well-known cognitive biases such as the conjunction fallacy, the representative heuristic, or the recognition heuristic, and investigate their relation to rule length and plausibility.

2019 ◽  
Vol 11 (17) ◽  
pp. 2057 ◽  
Author(s):  
Majid Shadman Roodposhti ◽  
Arko Lucieer ◽  
Asim Anees ◽  
Brett Bryan

This paper assesses the performance of DoTRules—a dictionary of trusted rules—as a supervised rule-based ensemble framework based on the mean-shift segmentation for hyperspectral image classification. The proposed ensemble framework consists of multiple rule sets with rules constructed based on different class frequencies and sequences of occurrences. Shannon entropy was derived for assessing the uncertainty of every rule and the subsequent filtering of unreliable rules. DoTRules is not only a transparent approach for image classification but also a tool to map rule uncertainty, where rule uncertainty assessment can be applied as an estimate of classification accuracy prior to image classification. In this research, the proposed image classification framework is implemented using three world reference hyperspectral image datasets. We found that the overall accuracy of classification using the proposed ensemble framework was superior to state-of-the-art ensemble algorithms, as well as two non-ensemble algorithms, at multiple training sample sizes. We believe DoTRules can be applied more generally to the classification of discrete data such as hyperspectral satellite imagery products.


2014 ◽  
Vol 2014 ◽  
pp. 1-6
Author(s):  
Shasha Li ◽  
Zhongmei Zhou ◽  
Weiping Wang

The Internet of things (IOT) is a hot issue in recent years. It accumulates large amounts of data by IOT users, which is a great challenge to mining useful knowledge from IOT. Classification is an effective strategy which can predict the need of users in IOT. However, many traditional rule-based classifiers cannot guarantee that all instances can be covered by at least two classification rules. Thus, these algorithms cannot achieve high accuracy in some datasets. In this paper, we propose a new rule-based classification, CDCR-P (Classification based on the Pruning and Double Covered Rule sets). CDCR-P can induce two different rule setsAandB. Every instance in training set can be covered by at least one rule not only in rule setA, but also in rule setB. In order to improve the quality of rule setB, we take measure to prune the length of rules in rule setB. Our experimental results indicate that, CDCR-P not only is feasible, but also it can achieve high accuracy.


2009 ◽  
Vol 18 (01) ◽  
pp. 1-16 ◽  
Author(s):  
RAMIN HALAVATI ◽  
SAEED BAGHERI SHOURAKI ◽  
SIMA LOTFI ◽  
POOYA ESFANDIAR

Evolutionary Algorithms are vastly used in development of rule based classifier systems in data mining where the rule base is usually a set of If-Then rules and an evolutionary trait develops and optimizes these rules. Genetic Algorithm is usually a favorite solution for such tasks as it globally searches for good rule-sets without any prior bias or greedy force, but it is usually slow. Also, designing a good genetic algorithm for rule base evolution requires the design of a recombination operator that merges two rule bases without disrupting the functionalities of each of them. To overcome the speed problem and the need to design recombination operator, this paper presents a novel algorithm for rule base evolution based on natural process of symbiogenesis. The algorithm uses symbiotic combination operator instead of traditional sexual recombination operator of genetic algorithms. This operator takes two chromosomes with different number of genes (rules here) and merges them by combining all the information content of both chromosomes. Using this operator results in two major advantages: First, it totally removes the need to design the recombination operator and therefore is easier to use; second, it outperforms traditional genetic algorithm both in emergence speed and classification rate, this is tested and presented on some globally used benchmarks.


2020 ◽  
Vol 20 (3-4) ◽  
pp. 218-237
Author(s):  
Oleg Sobchuk ◽  
Peeter Tinits

Abstract In many films, story is presented in an order different from chronological. Deviations from the chronological order in a narrative are called anachronies. Narratological theory and the evidence from psychological experiments indicate that anachronies allow stories to be more interesting, as the non-chronological order evokes curiosity in viewers. In this paper we investigate the historical dynamics in the use of anachronies in film. Particularly, we follow the cultural attraction theory that suggests that, given certain conditions, cultural evolution should conform to our cognitive preferences. We study this on a corpus of 80 most popular mystery films released in 1970–2009. We observe that anachronies have become used more frequently, and in a greater proportion of films. We also find that films that made substantial use of anachronies, on average, distributed the anachronies evenly along film length, while the films that made little use of anachronies placed them near the beginning and end. We argue that this can reflect a functional difference between these two types of using anachronies. The paper adds further support to the argument that popular culture may be influenced to a significant degree by our cognitive biases.


2019 ◽  
Vol 6 (2) ◽  
pp. 89-108 ◽  
Author(s):  
Stylianos Zikos ◽  
Maria Tsourma ◽  
Evdoxia E. Lithoxoidou ◽  
Anastasios Drosou ◽  
Dimosthenis Ioannidis ◽  
...  

This study evaluates user acceptance of a gamification-enabled collaboration and knowledge sharing platform that has been developed for use by personnel in industrial work environments, aiming at increasing motivation for knowledge exchange. The platform has been evaluated at two manufacturing industries by two groups of users, workers and supervisors, with regard to five criteria: usability, knowledge integration, working experience, user acceptance and overall impact. Results showed that even though the ratings from both industries were positive on all criteria, there is room for improvement on user acceptance and knowledge integration. Driven by this fact, a rule-based adaptive gamification approach which exploits information about workers is proposed in order to further increase motivation and engagement. Based on feedback received from the evaluation, guidelines related to functionalities and design of a gamified collaboration platform are provided. These guidelines can be followed when implementing collaboration tools with gamification support for industrial environments.


2020 ◽  
Vol 10 (4) ◽  
pp. 271-285
Author(s):  
Janusz T. Starczewski ◽  
Piotr Goetzen ◽  
Christian Napoli

AbstractIn real-world approximation problems, precise input data are economically expensive. Therefore, fuzzy methods devoted to uncertain data are in the focus of current research. Consequently, a method based on fuzzy-rough sets for fuzzification of inputs in a rule-based fuzzy system is discussed in this paper. A triangular membership function is applied to describe the nature of imprecision in data. Firstly, triangular fuzzy partitions are introduced to approximate common antecedent fuzzy rule sets. As a consequence of the proposed method, we obtain a structure of a general (non-interval) type-2 fuzzy logic system in which secondary membership functions are cropped triangular. Then, the possibility of applying so-called regular triangular norms is discussed. Finally, an experimental system constructed on precise data, which is then transformed and verified for uncertain data, is provided to demonstrate its basic properties.


Author(s):  
Carlos Pinheiro ◽  
Fernando Gomide ◽  
Otávio Carpinteiro ◽  
Isaías Lima

This chapter suggests a new method to develop rule-based models using concepts about rough sets. The rules encapsulate relations among variables and give a mechanism to link granular descriptions of the models with their computational procedures. An estimation procedure is suggested to compute values from granular representations encoded by rule sets. The method is useful to develop granular models of static and dynamic nonlinear systems and processes. Numerical examples illustrate the main features and the usefulness of the method.


2019 ◽  
Vol 37 (15_suppl) ◽  
pp. 3066-3066
Author(s):  
Yuan Ji ◽  
Meizi Liu

3066 Background: Other than the 3+3 design, new model-based statistical designs like the mTPI design (Ji and Wang, 2013, JCO) are alternative choices for oncology dose-finding trials, including immune oncology dose-finding trials (Atkins et al., 2018, Lancet Oncology). One major criticism of the 3+3 design is that it is based on simple rules, does not depend on statistical models for inference, and leads to unsafe and unreliable operating characteristics. However, the rule-based nature allows 3+3 to be easily understood and implemented in practice, making it practically attractive and friendly. Can friendly rule-based designs achieve great performance seen in model-based designs? For four decades, the answer has been NO. Methods: We propose a new rule-based design called i3+3, where the letter "i" represents the word "interval". The i3+3 design is based on simple but more clever rules that account for the variabilities in the observed data. In short, the i3+3 design simply asks clinicians to compare observed toxicity rates with a prespecified toxicity interval, and make dose escalation decisions according to three simple rules. No sophisticated modeling is needed and the entire design is transparent to clinicians. Results: We compare the operating characteristics for the proposed i3+3 design with other popular phase I designs by simulation. The i3+3 design is far superior than the 3+3 design in trial safety and the ability to identify the true MTD. Compared with model-based phase I designs, i3+3 also demonstrates comparable performances. In other words, the i3+3 design possesses both simplicity and transparency of the rule-based approaches, and the superior operating characteristics seen in model-based approaches. An online R Shiny tool is provided to illustrate the i3+3 design, although in practice it requires no software to design or conduct a dose-finding trial using the design. Conclusions: The i3+3 design could be a practice-altering method for the clinical community. It may increase the safety and efficiency of dose finding trials.


AI ◽  
2021 ◽  
Vol 2 (4) ◽  
pp. 621-635
Author(s):  
Vincent Margot ◽  
George Luta

Interpretability is becoming increasingly important for predictive model analysis. Unfortunately, as remarked by many authors, there is still no consensus regarding this notion. The goal of this paper is to propose the definition of a score that allows for quickly comparing interpretable algorithms. This definition consists of three terms, each one being quantitatively measured with a simple formula: predictivity, stability and simplicity. While predictivity has been extensively studied to measure the accuracy of predictive algorithms, stability is based on the Dice-Sorensen index for comparing two rule sets generated by an algorithm using two independent samples. The simplicity is based on the sum of the lengths of the rules derived from the predictive model. The proposed score is a weighted sum of the three terms mentioned above. We use this score to compare the interpretability of a set of rule-based algorithms and tree-based algorithms for the regression case and for the classification case.


Author(s):  
Vicente Arturo Romero Zaldivar ◽  
Daniel Burgos ◽  
Abelardo Pardo

Recommendation Systems are central in current applications to help the user find relevant information spread in large amounts of data. Most Recommendation Systems are more effective when huge amounts of user data are available. Educational applications are not popular enough to generate large amount of data. In this context, rule-based Recommendation Systems seem a better solution. Rules can offer specific recommendations with even no usage information. However, large rule-sets are hard to maintain, reengineer, and adapt to user preferences. Meta-rules can generalize a rule-set which provides bases for adaptation. In this chapter, the authors present the benefits of meta-rules, implemented as part of Meta-Mender, a meta-rule based Recommendation System. This is an effective solution to provide a personalized recommendation to the learner, and constitutes a new approach to Recommendation Systems.


Sign in / Sign up

Export Citation Format

Share Document