Scale development with small samples: a new application of longitudinal item response theory

2018 ◽  
Vol 27 (7) ◽  
pp. 1721-1734 ◽  
Author(s):  
Carrie R. Houts ◽  
Robert Morlock ◽  
Steven I. Blum ◽  
Michael C. Edwards ◽  
R. J. Wirth
2019 ◽  
Vol 38 (6) ◽  
pp. 692-705 ◽  
Author(s):  
Boris Forthmann ◽  
Rüdiger Grotjahn ◽  
Philipp Doebler ◽  
Purya Baghaei

As measures of general language proficiency, C-tests are ubiquitous in language testing. Speeded C-tests are quite recent developments in the field and are deemed to be more discriminatory and provide more accurate diagnostic information than power C-tests especially with high-ability participants. Item response theory modeling of speeded C-tests has not been discussed in the literature, and current approaches for power C-tests based on ordinal models either violate the model assumptions or are relatively complex to be reliably fitted with small samples. Count data models are viable alternatives with less restrictive assumptions and lower complexity. In the current study, we compare count data models with commonly applied ordinal models for modeling a speeded C-test. It was found that a flexible count data model fits equally well in absolute and relative terms as compared with ordinal models. Implications and feasibility of count data models for the psychometric modeling of C-tests are discussed.


2021 ◽  
Author(s):  
Jakob Bue Bjorner ◽  
Berend Terluin ◽  
Andrew Trigg ◽  
Jinxiang Hu ◽  
Keri J.S. Brady ◽  
...  

Abstract PURPOSE: Thresholds for meaningful within-individual change (MWIC) are useful for interpreting patient-reported outcome measures (PROM). Transition ratings (TR) have been recommended as anchors to establish MWIC. Traditional statistical methods for analyzing MWIC such as mean change analysis, receiver operating characteristic (ROC) analysis, and predictive modeling ignore problems of floor/ceiling effects and measurement error in the PROM scores and the TR item. We present a novel approach to MWIC estimation for multi-item scales using longitudinal item response theory (LIRT).METHODS: A Graded Response LIRT model for baseline and follow-up PROM data was expanded to include a TR item measuring latent change. The LIRT threshold parameter for the TR established the MWIC threshold on the latent metric, from which the observed PROM score MWIC threshold was estimated. We compared the LIRT approach and traditional methods using an example data set with baseline and three follow-up assessments differing by magnitude of score improvement, variance of score improvement, and baseline-follow-up score correlation.RESULTS: The LIRT model provided good fit to the data. LIRT estimates of observed PROM MWIC varied between 3 and 4 points score improvement. In contrast, results from traditional methods varied from 2 points to 10 points - strongly associated with proportion of self-rated improvement. Best agreement between methods was seen when approximately 50% rated their health as improved.CONCLUSION : Results from traditional analyses of anchor-based MWIC are impacted by study conditions. LIRT constitutes a promising and more robust analytic approach to identifying thresholds for MWIC.


2014 ◽  
Vol 10 ◽  
pp. P910-P910
Author(s):  
Arlette Edjolo ◽  
Cécile Proust-Lima ◽  
Catherine Helmer ◽  
Fleur Delva ◽  
Jean-Francois Dartigues ◽  
...  

2016 ◽  
Vol 44 (2) ◽  
pp. 216-225 ◽  
Author(s):  
Tiffany A. Whittaker ◽  
Roger L. Worthington

1999 ◽  
Vol 48 (4) ◽  
pp. 234-238 ◽  
Author(s):  
Miaofen Yen ◽  
Sandra R. Edwardson

Sign in / Sign up

Export Citation Format

Share Document