Can Serious Games Assess Decision-Making Biases?

2020 ◽  
Vol 36 (1) ◽  
pp. 44-55
Author(s):  
Kyoungwon Seo ◽  
Hokyoung Ryu ◽  
Jieun Kim

Abstract. The limitations of self-report questionnaires and interview methods for assessing individual differences in human cognitive biases have become increasingly apparent. These limitations have led to a renewed interest in alternative modes of assessment, including for implicit and explicit aspects of human behavior (i.e., dual-process theory). Acknowledging this, the present study was conducted to develop and validate a serious game, “Don Quixote,” for measuring specific cognitive biases: the bandwagon effect and optimism bias. We hypothesized that the implicit and explicit game data would mirror the results from an interview and questionnaire, respectively. To examine this hypothesis, participants ( n = 135) played the serious game and completed a questionnaire and interview in a random order for cross-validation. The results demonstrated that the implicit game data (e.g., response time) were highly correlated with the interview data. On the contrary, the explicit game data (e.g., game score) were comparable to the results from the questionnaire. These findings suggest that the serious game and the underlying intrinsic nature of its game mechanics (i.e., evoking instant responses under time pressure) are of importance for the further development of cognitive bias measures in both academia and practice.

Author(s):  
Brian A. Nosek ◽  
Frederick L. Smyth

Abstract. Recent theoretical and methodological innovations suggest a distinction between implicit and explicit evaluations. We applied Campbell and Fiske's (1959) classic multitrait-multimethod design precepts to test the construct validity of implicit attitudes as measured by the Implicit Association Test (IAT). Participants (N = 287) were measured on both self-report and IAT for up to seven attitude domains. Through a sequence of latent-variable structural models, systematic method variance was distinguished from attitude variance, and a correlated two-factors-per-attitude model (implicit and explicit factors) was superior to a single-factor-per-attitude specification. That is, despite sometimes strong relations between implicit and explicit attitude factors, collapsing their indicators into a single attitude factor resulted in relatively inferior model fit. We conclude that these implicit and explicit measures assess related but distinct attitude constructs. This provides a basis for, but does not distinguish between, dual-process and dual-representation theories that account for the distinctions between constructs.


Diagnosis ◽  
2020 ◽  
Vol 0 (0) ◽  
Author(s):  
Sumner Abraham ◽  
Andrew Parsons ◽  
Brian Uthlaut ◽  
Peggy Plews-Ogan

AbstractDespite the breadth of patient safety initiatives, physicians talking about their mistakes to other physicians is a difficult thing to do. This difficulty may be exacerbated by a limited exposure to how to analyze and discuss mistakes and respond in a productive way. At the University of Virginia, we recognized the importance of understanding cognitive biases for residents in both their clinical and personal professional development. We re-designed our resident led morbidity and mortality (M&M) conference using a model that integrates dual-process theory and metacognition to promote informed reflection and analysis of cognitive diagnostic errors. We believe that structuring M&M in this way builds a culture that encourages reflection together to learn our most difficult diagnostic errors and to engage in where our thought processes went wrong. In slowly building this culture, we hope to inoculate residents with the habits of mind that can best protect them from harmful biases in their clinical reasoning while instilling a culture of self-reflection.


Author(s):  
Eliezer Yudkowsky

By far the greatest danger of Artificial Intelligence (AI) is that people conclude too early that they understand it. Of course, this problem is not limited to the field of AI. Jacques Monod wrote: ‘A curious aspect of the theory of evolution is that everybody thinks he understands it’ (Monod, 1974). The problem seems to be unusually acute in Artificial Intelligence. The field of AI has a reputation for making huge promises and then failing to deliver on them. Most observers conclude that AI is hard, as indeed it is. But the embarrassment does not stem from the difficulty. It is difficult to build a star from hydrogen, but the field of stellar astronomy does not have a terrible reputation for promising to build stars and then failing. The critical inference is not that AI is hard, but that, for some reason, it is very easy for people to think they know far more about AI than they actually do. It may be tempting to ignore Artificial Intelligence because, of all the global risks discussed in this book, AI is probably hardest to discuss. We cannot consult actuarial statistics to assign small annual probabilities of catastrophe, as with asteroid strikes. We cannot use calculations from a precise, precisely confirmed model to rule out events or place infinitesimal upper bounds on their probability, as with proposed physics disasters. But this makes AI catastrophes more worrisome, not less. The effect of many cognitive biases has been found to increase with time pressure, cognitive busyness, or sparse information. Which is to say that the more difficult the analytic challenge, the more important it is to avoid or reduce bias. Therefore I strongly recommend reading my other chapter (Chapter 5) in this book before continuing with this chapter. When something is universal enough in our everyday lives, we take it for granted to the point of forgetting it exists. Imagine a complex biological adaptation with ten necessary parts. If each of the ten genes is independently at 50% frequency in the gene pool – each gene possessed by only half the organisms in that species – then, on average, only 1 in 1024 organisms will possess the full, functioning adaptation.


2021 ◽  
Vol 50 (Supplement_1) ◽  
Author(s):  
Jie Zhang ◽  
Anja Olsen ◽  
Jytte Halkjær ◽  
Kristina Elin Nielsen Petersen ◽  
Anne Tjønneland ◽  
...  

Abstract Background It is easy and cost-effective to ask study participants to self-report height and weight and self-reported anthropometry is therefore widely used in epidemiological studies. However, it is questioned to what degree self-reported adiposity indices are a solid proxy of measured indices in terms of estimates of health outcomes. The current study aimed to quantify the agreement between self-reported and measured anthropometrics, including height, weight, body mass index (BMI), weight circumference (WC), and weight-to-height ratio (WHtR) in a contemporary cohort of adults, and to assess whether anthropometric indices misreporting yielded inaccurate estimates of associations with cardiometabolic biomarkers. Methods Self-reported and measured anthropometric variables were obtained from the Diet, Cancer, and Health-Next Generation Cohort (n = 39,514). Pearson correlations and Lin’s concordance correlations evaluated the correlation between self-report and measured anthropometrics. Misreporting in relation to age, sex and smoking status was investigated. Multivariable regression models and ROC analyses were used to assess the associations of cardiometabolic biomarkers with self-reported and measured general obesity and abdominal obesity. Results Self-reported height was overreported by 1.07 cm, weight was underreported by 0.32 kg on average, which led to self-reported BMI 0.42 kg/m2 lower than measured. Self-reported and measured height, weight, BMI, WC and WHtR were highly correlated (r = 0.98, 0.99, 0.98, 0.88, 0.86, respectively). Associations between self-reported indices and cardiometabolic biomarkers were comparable to associations assessed with measured anthropometrics. Conclusions The self-reported anthropometric indices were reliable when estimating associations with metabolic biomarkers. Key messages This study found overall agreement between self-reported and measured anthropometric variables.


1969 ◽  
Vol 25 (3) ◽  
pp. 711-714 ◽  
Author(s):  
Russell G. Geen ◽  
Robert George

A self-report inventory made up of items from the Buss-Durkee manifest aggressiveness scales, the Marlowe-Crowne Social Desirability Scale, and the Masculinity-Femininity scale of the Guilford-Zimmerman Temperament Survey was administered to 72 men along with a test of verbal associations to aggressive and neutral cue words. The number of aggressive associations made to aggressive cue words was highly correlated with over-all manifest aggressiveness and with two of the aggressiveness subscales. The results were discussed in terms of the relationship of aggressiveness habit strength to verbal behavior.


Assessment ◽  
1996 ◽  
Vol 3 (1) ◽  
pp. 17-25 ◽  
Author(s):  
Dean Lauterbach ◽  
Scott Vrana

This paper describes three studies of the reliability and validity of a newly revised version of the Purdue Posttraumatic Stress Disorder scale (PPTSD-R). The PPTSD-R is a 17-item questionnaire that yields four scores: Reexperiencing, Avoidance, Arousal, and Total. It is highly internally consistent (α = .91), and the scores are relatively stable across time. The PPTSD-R is highly correlated with other measures of PTSD symptomatology and moderately correlated with measures of related psychopathology, providing preliminary support for the measure's convergent and discriminant validity. It reliably distinguishes between groups of people who were and were not traumatized, it is sensitive to the impact of different types of traumatic events, and (within a clinical sample) it discriminates between those who did and did not seek treatment for difficulty coping with the traumatic event being assessed. The PPTSD-R shows promise as a measure of PTSD symptoms in the college population.


Diagnosis ◽  
2018 ◽  
Vol 5 (1) ◽  
pp. 11-14 ◽  
Author(s):  
Robert L. Trowbridge ◽  
Andrew P.J. Olson

AbstractDiagnostic reasoning is one of the most challenging and rewarding aspects of clinical practice. As a result, facility in teaching diagnostic reasoning is a core necessity for all medical educators. Clinician educators’ limited understanding of the diagnostic process and how expertise is developed may result in lost opportunities in nurturing the diagnostic abilities of themselves and their learners. In this perspective, the authors describe their journeys as clinician educators searching for a coherent means of teaching diagnostic reasoning. They discuss the initial appeal and immediate applicability of dual process theory and cognitive biases to their own clinical experiences and those of their trainees, followed by the eventual and somewhat belated recognition of the importance of context specificity. They conclude that there are no quick fixes in guiding learners to expertise of diagnostic reasoning, but rather the development of these abilities is best viewed as a long, somewhat frustrating, but always interesting journey. The role of the teacher of clinical reasoning is to guide the learners on this journey, recognizing true mastery may not be attained, but should remain a goal for teacher and learner alike.


Assessment ◽  
2021 ◽  
pp. 107319112110556
Author(s):  
Stephen L. Aita ◽  
Grant G. Moncrief ◽  
Jennifer Greene ◽  
Sue Trujillo ◽  
Alicia Carrillo ◽  
...  

The Behavior Rating Inventory of Executive Function–Adult Version (BRIEF-A) is a standardized rating scale of subjective executive functioning. We provide univariate and multivariate base rates (BRs) for scale/index scores in the clinical range ( T scores ≥65), reliable change, and inter-rater information not included in the Professional Manual. Participants were adults (ages = 18–90 years) from the BRIEF-A self-report ( N = 1,050) and informant report ( N = 1,200) standardization samples, as well as test–retest ( n = 50 for self, n = 44 for informant) and inter-rater ( n = 180) samples. Univariate BRs of elevated T scores were low (self-report = 3.3%–15.4%, informant report = 4.5%–16.3%). Multivariate BRs revealed the common occurrence of obtaining at least one elevated T-score across scales (self-report = 26.5%–37.3%, informant report = 22.7%–30.3%), whereas virtually none had elevated scores on all scales. Test–retest scores were highly correlated (self = .82–.94; informant = .91–.96). Inter-rater correlations ranged from .44 to .68. Significant ( p < .05) test–retest T-score differences ranged from 7 to 12 for self-report, from 6 to 8 for informant report, and from 16 to 21 points for inter-rater T-score differences. Applications of these findings are discussed.


2017 ◽  
pp. 1592-1612
Author(s):  
Gwendolyn A. Kelso ◽  
Leslie R. Brody

Stereotype threat about leadership ability may trigger emotional and cognitive responses that reduce women's leadership aspirations. This chapter reviews literature and presents a study on the effects of implicit (covert) and explicit (overt) leadership stereotype threat on women's emotions, power-related cognitions, and behaviors as moderated by exposure to powerful female or male role models. Emotional responses were measured using self-report (direct) and narrative writing (indirect) tasks. Undergraduate women (n = 126) in the Northeastern U.S. were randomly divided into three stereotype threat groups: none, implicit, and explicit. Implicit stereotype threat resulted in higher indirectly expressed (but not self-reported) anxiety, behaviors that benefited others more than the self, and when preceded by exposure to powerful female role models, higher self-reported negative emotion but also higher indirect positive affect. Explicit stereotype threat resulted in higher indirect optimism, and when preceded by exposure to powerful female role models, lower self-reported sadness but also lower implicit power cognitions.


2019 ◽  
Vol 37 (1) ◽  
pp. 27-46
Author(s):  
Laura M. River ◽  
Angela J. Narayan ◽  
Victoria M. Atzl ◽  
Luisa M. Rivera ◽  
Alicia F. Lieberman

Romantic partner support from the father-to-be is associated with women’s mental health during pregnancy. However, most studies of partner support rely upon women’s responses to self-report questionnaires, which may be biased and should be corroborated by efficient, coder-rated measures of partner support. This study tested whether the Five-Minute Speech Sample (FMSS), adapted to assess expressed emotion about romantic partners, can provide information about partner support during pregnancy that is less prone to bias than self-report. Participants were 101 low-income, ethnically diverse pregnant women who completed self-report questions on partner support quality and the FMSS. Self-reported and coder-rated (FMSS) partner support were highly correlated and were each significantly associated with self-reported depressive and post-traumatic stress disorder (PTSD) symptoms, perceived stress, and partner victimization during pregnancy. Self-reported and coder-rated support corresponded in approximately 75% of cases; however, nearly 25% of women self-reported high support but received low FMSS support ratings. These women reported elevated PTSD symptoms, perceived stress, and victimization during pregnancy. While self-reported partner support may be valid for many respondents, the FMSS is less susceptible to reporting biases and may better identify women facing heightened psychopathology and stress during pregnancy, who would benefit from supportive intervention.


Sign in / Sign up

Export Citation Format

Share Document