scholarly journals Test-retest reliability on the Cambridge Neuropsychological Test Automated Battery: Comment on Karlsen et al. (2020)

Author(s):  
Caroline Skirrow ◽  
Nathan Cashdollar ◽  
Kiri Granger ◽  
Sally Jennings ◽  
Elizabeth Baker ◽  
...  
2019 ◽  
Vol 44 (6) ◽  
pp. 365-369
Author(s):  
Davangere P Devanand ◽  
Xinhua Liu ◽  
Hannah Cohen ◽  
John Budrow ◽  
Nicole Schupf ◽  
...  

Abstract The objective of this study was to determine the long-term test–retest reliability of the University of Pennsylvania Smell Identification Test (UPSIT), and its individual items, in cognitively intact older adults. A community sample of older adults received a neuropsychological test battery, including the 12-item, 6-trial Selective Reminding Test (SRT). The UPSIT was administered at baseline and follow-up that occurred between 1 and 4 years after baseline. UPSIT scores of participants who were cognitively intact and did not decline cognitively were examined for test–retest reliability. In 92 older adults with mean age 77.6 years followed for 2.79 (standard deviation [SD] 0.69) years, mean UPSIT score declined from 30.29 (SD 5.83) to 27.80 (SD 5.50). In linear mixed models that adjusted for time, age, sex, and education, intraclass correlation coefficients for UPSIT were 0.65, SRT delayed recall 0.59, and SRT total immediate recall 0.49. Among 4 possible response combinations, the largest proportion of participants had correct responses at both visits for 35 out of 40 items. Consistency of item responses ranged from 50% to 90% across the 2 time points. The long-term test–retest reliability of the UPSIT was moderately strong without practice effects over long periods of time in older adults. These results provide indirect support to prior findings on odor identification impairment predicting cognitive decline and dementia, and suggest potential use of olfactory testing as a biomarker in prevention and treatment trials of cognitive enhancers.


Author(s):  
Rune H. Karlsen ◽  
Justin E. Karr ◽  
Simen B. Saksvik ◽  
Astri J. Lundervold ◽  
Odin Hjemdal ◽  
...  

2021 ◽  
Author(s):  
Kelly L. Sloane ◽  
Joel A. Mefford ◽  
Shenly Glenn ◽  
Zilong Zhao ◽  
Man Xu ◽  
...  

AbstractBackgroundA mobile platform for the self-administration of sensor-based cognitive and behavioral assessments was developed. In addition to measurements typical of legacy neuropsychological tests, Miro also quantifies functions that are currently left to subjective clinical impressions such as motor function, language, and speech.ObjectiveStudies were conducted to measure Miro’s concurrent validity, test-retest reliability, and amnestic MCI classification performance.MethodSpearman correlations were calculated to estimate the concurrent validity of Miro and legacy variables using data from 207 study participants. Twenty-six healthy controls were assessed at three time points to evaluate the test-retest reliability of Miro test scores. Reliability was quantified with the scores’ intraclass correlations. Learning effects were measured as trends over three assessments. A machine learning algorithm combined Miro scores into a Risk Score to distinguish 65 healthy controls, 21 amnestic MCI (aMCI) participants, and 17 non-amnestic MCI (naMCI) participants.ResultsSignificant correlations of Miro variables with legacy neuropsychological test variables were observed. Longitudinal studies show agreement of subsequent measurements and minimal learning effects. The Risk Score distinguished aMCI from healthy controls with an Area Under the Receiver Operator Curve (AUROC) of 0.97; the naMCI participants and controls were separated with an AUROC of 0.80, and the combined (aMCI + naMCI) group was separated from healthy controls with an AUROC of 0.89.ConclusionMiro includes valid and reliable versions of scores from legacy neuropsychological test scores and a machine-learning derived risk score that effectively distinguishes healthy controls and individuals with MCI.


Author(s):  
Matthew L. Hall ◽  
Stephanie De Anda

Purpose The purposes of this study were (a) to introduce “language access profiles” as a viable alternative construct to “communication mode” for describing experience with language input during early childhood for deaf and hard-of-hearing (DHH) children; (b) to describe the development of a new tool for measuring DHH children's language access profiles during infancy and toddlerhood; and (c) to evaluate the novelty, reliability, and validity of this tool. Method We adapted an existing retrospective parent report measure of early language experience (the Language Exposure Assessment Tool) to make it suitable for use with DHH populations. We administered the adapted instrument (DHH Language Exposure Assessment Tool [D-LEAT]) to the caregivers of 105 DHH children aged 12 years and younger. To measure convergent validity, we also administered another novel instrument: the Language Access Profile Tool. To measure test–retest reliability, half of the participants were interviewed again after 1 month. We identified groups of children with similar language access profiles by using hierarchical cluster analysis. Results The D-LEAT revealed DHH children's diverse experiences with access to language during infancy and toddlerhood. Cluster analysis groupings were markedly different from those derived from more traditional grouping rules (e.g., communication modes). Test–retest reliability was good, especially for the same-interviewer condition. Content, convergent, and face validity were strong. Conclusions To optimize DHH children's developmental potential, stakeholders who work at the individual and population levels would benefit from replacing communication mode with language access profiles. The D-LEAT is the first tool that aims to measure this novel construct. Despite limitations that future work aims to address, the present results demonstrate that the D-LEAT represents progress over the status quo.


1982 ◽  
Vol 25 (4) ◽  
pp. 521-527 ◽  
Author(s):  
David C. Shepherd

In 1977, Shepherd and colleagues reported significant correlations (–.90, –.91) between speechreading scores and the latency of a selected negative peak (VN 130 measure) on the averaged visual electroencephalic wave form. The primary purpose of this current study was to examine the stability, or repeatability, of this relation between these cognitive and neurophysiologic measures over a period of several months and thus support its test-retest reliability. Repeated speechreading word and sentence scores were gathered during three test-retest sessions from each of 20 normal-hearing adults. An average of 56 days occurred from the end of one to the beginning of another speechreading sessions. During each of four other test-retest sessions, averaged visual electroencephalic responses (AVER s ) were evoked from each subject. An average of 49 clays intervened between AVER sessions. Product-moment correlations computed among repeated word scores and VN l30 measures ranged from –.61 to –.89. Based on these findings, it was concluded that the VN l30 measure of visual neural firing time is a reliable correlate of speech-reading in normal-hearing adults.


2000 ◽  
Vol 16 (1) ◽  
pp. 53-58 ◽  
Author(s):  
Hans Ottosson ◽  
Martin Grann ◽  
Gunnar Kullgren

Summary: Short-term stability or test-retest reliability of self-reported personality traits is likely to be biased if the respondent is affected by a depressive or anxiety state. However, in some studies, DSM-oriented self-reported instruments have proved to be reasonably stable in the short term, regardless of co-occurring depressive or anxiety disorders. In the present study, we examined the short-term test-retest reliability of a new self-report questionnaire for personality disorder diagnosis (DIP-Q) on a clinical sample of 30 individuals, having either a depressive, an anxiety, or no axis-I disorder. Test-retest scorings from subjects with depressive disorders were mostly unstable, with a significant change in fulfilled criteria between entry and retest for three out of ten personality disorders: borderline, avoidant and obsessive-compulsive personality disorder. Scorings from subjects with anxiety disorders were unstable only for cluster C and dependent personality disorder items. In the absence of co-morbid depressive or anxiety disorders, mean dimensional scores of DIP-Q showed no significant differences between entry and retest. Overall, the effect from state on trait scorings was moderate, and it is concluded that test-retest reliability for DIP-Q is acceptable.


Sign in / Sign up

Export Citation Format

Share Document