scholarly journals Comparing five depression measures in depressed Chinese patients using item response theory: an examination of item properties, measurement precision and score comparability

2017 ◽  
Vol 15 (1) ◽  
Author(s):  
Yue Zhao ◽  
Wai Chan ◽  
Barbara Chuen Yee Lo
Assessment ◽  
2019 ◽  
Vol 27 (7) ◽  
pp. 1416-1428 ◽  
Author(s):  
Hao Luo ◽  
Björn Andersson ◽  
Jennifer Y. M. Tang ◽  
Gloria H. Y. Wong

The traditional application of the Montreal Cognitive Assessment uses total scores in defining cognitive impairment levels, without considering variations in item properties across populations. Item response theory (IRT) analysis provides a potential solution to minimize the effect of important confounding factors such as education. This research applies IRT to investigate the characteristics of Montreal Cognitive Assessment items in a randomly selected, culturally homogeneous sample of 1,873 older persons with diverse educational backgrounds. Any formal education was used as a grouping variable to estimate multiple-group IRT models. Results showed that item characteristics differed between people with and without formal education. Item functioning of the Cube, Clock Number, and Clock Hand items was superior in people without formal education. This analysis provided evidence that item properties vary with education, calling for more sophisticated modelling based on IRT to incorporate the effect of education.


Author(s):  
Anju Devianee Keetharuth ◽  
Jakob Bue Bjorner ◽  
Michael Barkham ◽  
John Browne ◽  
Tim Croudace ◽  
...  

Abstract Purpose ReQoL-10 and ReQoL-20 have been developed for use as outcome measures with individuals aged 16 and over, experiencing mental health difficulties. This paper reports modelling results from the item response theory (IRT) analyses that were used for item reduction. Methods From several stages of preparatory work including focus groups and a previous psychometric survey, a pool of items was developed. After confirming that the ReQoL item pool was sufficiently unidimensional for scoring, IRT model parameters were estimated using Samejima’s Graded Response Model (GRM). All 39 mental health items were evaluated with respect to item fit and differential item function regarding age, gender, ethnicity, and diagnosis. Scales were evaluated regarding overall measurement precision and known-groups validity (by care setting type and self-rating of overall mental health). Results The study recruited 4266 participants with a wide range of mental health diagnoses from multiple settings. The IRT parameters demonstrated excellent coverage of the latent construct with the centres of item information functions ranging from − 0.98 to 0.21 and with discrimination slope parameters from 1.4 to 3.6. We identified only two poorly fitting items and no evidence of differential item functioning of concern. Scales showed excellent measurement precision and known-groups validity. Conclusion The results from the IRT analyses confirm the robust structure properties and internal construct validity of the ReQoL instruments. The strong psychometric evidence generated guided item selection for the final versions of the ReQoL measures.


2011 ◽  
Vol 20 (3) ◽  
pp. e39-e55 ◽  
Author(s):  
Muirne C. S. Paap ◽  
Rob R. Meijer ◽  
Jan Van Bebber ◽  
Geir Pedersen ◽  
Sigmund Karterud ◽  
...  

2003 ◽  
Vol 25 ◽  
pp. D34-D36 ◽  
Author(s):  
Jin-she Lai ◽  
Kelly Dineen ◽  
David Cella ◽  
Jamie von Roenn

Author(s):  
Milena D Anatchkova ◽  
Constance Barysauskas ◽  
Rebecca L Kinney ◽  
Lisa Lombardini ◽  
Jeroan J Allison

Objective: Quality of transitional care is associated with important health outcomes such as rehospitalization, increased medical errors, and costs. A widely used measure of the construct, the Care Transitions Measure (CTM-15), was developed and validated with relatively small samples and classical test theory approach. The short version of the measure (CTM-3) was recently included in the CAHPS® Hospital Survey. We aimed to conduct a psychometric evaluation of the CTM-15 in a large sample and evaluate whether an item response theory (IRT) based scoring would lead to better measurement precision. Methods: As part of TRACE-CORE 1545 participants were interviewed during hospitalization for ACS providing information on general health status (SF-36). At 1 month following discharge, patients completed CTM items, health utilization and care process questions. We evaluated the psychometric properties of the CTM using descriptive statistics, factor analyses, and item response theory analyses. We compared the measurement precision of CTM-15, CTM-3, and a CTM-IRT based score using relative validity (RV) coefficients. Results: Participants were 79% non-Hispanic white, 67% male, 27% with a college education or higher (27%) and average age of 62 years. The CTM scale had good internal consistency (Cronbach’s alpha=0.95), but demonstrated strong acquiescence bias (8.7% participants responded “Strongly agree”, 19% “Agree” to all 15 items) and limited score variability (5% <50score). A bi-factor model with one general factor with high item loadings and four secondary factors with lower item loadings, fit the data best supporting the assumption of essential unidimensionality of the measure. Review of item characteristic curves indicated presence of non-discriminating response categories in all items. IRT based item parameters were estimated for all items, but for five of the items there was evidence for model misfit. The CTM-15 differentiated between groups of patients defined by self-reported health status, health care utilization, and care transition process indicators. Differences between groups were small (2-3 points). There was no gain in measurement precision for the scale from IRT scoring. The CTM-3 was not significantly lower for patients reporting rehospitalization or emergency department visits. Conclusion: We identified psychometric challenges of the CTM, which may limit its value in research and practice. The strong acquiescence bias in the measure leads to highly skewed, clustered scores with restricted score variance and makes it difficult to differentiate between levels of care transition quality. In the absence of guidelines on meaningfully important differences, it is hard to determine whether detected statistically significant differences in care transition quality are important. These results are in line with emerging evidence of gaps in the validity of the measure.


2005 ◽  
Vol 30 (3) ◽  
pp. 278-288 ◽  
Author(s):  
Jin-shei Lai ◽  
Kelly Dineen ◽  
Bryce B. Reeve ◽  
Jamie Von Roenn ◽  
Daniel Shervin ◽  
...  

2016 ◽  
Vol 59 (2) ◽  
pp. 281-289 ◽  
Author(s):  
Guido Makransky ◽  
Philip S. Dale ◽  
Philip Havmose ◽  
Dorthe Bleses

Purpose This study investigated the feasibility and potential validity of an item response theory (IRT)–based computerized adaptive testing (CAT) version of the MacArthur–Bates Communicative Development Inventory: Words & Sentences (CDI:WS; Fenson et al., 2007) vocabulary checklist, with the objective of reducing length while maintaining measurement precision. Method Parent-reported vocabulary for the American CDI:WS norming sample consisting of 1,461 children between the ages of 16 and 30 months was used to investigate the fit of the items to the 2-parameter logistic IRT model and to simulate CDI-CAT versions with 400, 200, 100, 50, 25, 10, and 5 items. Results All but 14 items fit the 2-parameter logistic IRT model, and real data simulations of CDI-CATs with at least 50 items recovered full CDI scores with correlations over .95. Furthermore, the CDI-CATs with at least 50 items had similar correlations with age and socioeconomic status as the full CDI:WS. Conclusion These results provide strong evidence that a CAT version of the CDI:WS has the potential to reduce length while maintaining the accuracy and precision of the full instrument.


2019 ◽  
Vol 80 (4) ◽  
pp. 726-755 ◽  
Author(s):  
Jinho Kim ◽  
Mark Wilson

This study investigates polytomous item explanatory item response theory models under the multivariate generalized linear mixed modeling framework, using the linear logistic test model approach. Building on the original ideas of the many-facet Rasch model and the linear partial credit model, a polytomous Rasch model is extended to the item location explanatory many-facet Rasch model and the step difficulty explanatory linear partial credit model. To demonstrate the practical differences between the two polytomous item explanatory approaches, two empirical studies examine how item properties explain and predict the overall item difficulties or the step difficulties each in the Carbon Cycle assessment data and in the Verbal Aggression data. The results suggest that the two polytomous item explanatory models are methodologically and practically different in terms of (a) the target difficulty parameters of polytomous items, which are explained by item properties; (b) the types of predictors for the item properties incorporated into the design matrix; and (c) the types of item property effects. The potentials and methodological advantages of item explanatory modeling are discussed as well.


Sign in / Sign up

Export Citation Format

Share Document