Using Rasch Analysis to Inform Rating Scale Development

2017 ◽  
Vol 58 (8) ◽  
pp. 922-933 ◽  
Author(s):  
Carol Van Zile-Tamsen
2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Mario Cantó-Cerdán ◽  
Pilar Cacho-Martínez ◽  
Francisco Lara-Lacárcel ◽  
Ángel García-Muñoz

AbstractTo develop the Symptom Questionnaire for Visual Dysfunctions (SQVD) and to perform a psychometric analysis using Rasch method to obtain an instrument which allows to detect the presence and frequency of visual symptoms related to any visual dysfunction. A pilot version of 33 items was carried out on a sample of 125 patients from an optometric clinic. Rasch model (using Andrich Rating Scale Model) was applied to investigate the category probability curves and Andrich thresholds, infit and outfit mean square, local dependency using Yen’s Q3 statistic, Differential item functioning (DIF) for gender and presbyopia, person and item reliability, unidimensionality, targeting and ordinal to interval conversion table. Category probability curves suggested to collapse a response category. Rasch analysis reduced the questionnaire from 33 to 14 items. The final SQVD showed that 14 items fit to the model without local dependency and no significant DIF for gender and presbyopia. Person reliability was satisfactory (0.81). The first contrast of the residual was 1.908 eigenvalue, showing unidimensionality and targeting was − 1.59 logits. In general, the SQVD is a well-structured tool which shows that data adequately fit the Rasch model, with adequate psychometric properties, making it a reliable and valid instrument to measure visual symptoms.


2021 ◽  
pp. 026553222199405
Author(s):  
Ute Knoch ◽  
Bart Deygers ◽  
Apichat Khamboonruang

Rating scale development in the field of language assessment is often considered in dichotomous ways: It is assumed to be guided either by expert intuition or by drawing on performance data. Even though quite a few authors have argued that rating scale development is rarely so easily classifiable, this dyadic view has dominated language testing research for over a decade. In this paper we refine the dominant model of rating scale development by drawing on a corpus of 36 studies identified in a systematic review. We present a model showing the different sources of scale construct in the corpus. In the discussion, we argue that rating scale designers, just like test developers more broadly, need to start by determining the purpose of the test, the relevant policies that guide test development and score use, and the intended score use when considering the design choices available to them. These include considering the impact of such sources on the generalizability of the scores, the precision of the post-test predictions that can be made about test takers’ future performances and scoring reliability. The most important contributions of the model are that it gives rating scale developers a framework to consider prior to starting scale development and validation activities.


2016 ◽  
Vol 33 (1) ◽  
pp. 74 ◽  
Author(s):  
Alejandro Veas ◽  
Juan Luis Castejón ◽  
Raquel Gilar ◽  
Pablo Miñano

<p>The School Attitude Assessment Survey-Revised (SAAS-R) was developed by McCoach &amp; Siegle (2003b) and validated in Spain by Author (2014) using Classical Test Theory. The objective of the current research is to validate SAAS-R using multidimensional Rasch analysis. Data were collected from 1398 students attending different high schools. Principal Component Analysis supported the multidimensional SAAS-R. The item difficulty and person ability were calibrated along the same latent trait scale. 10 items were removed from the scale due to misfit with the Rasch model. Differential Item Functioning revealed no significant differences across gender for the remaining 25 items. The 7-category rating scale structure did not function well, and the subscale goal valuation obtained low reliability values. The multidimensional Rasch model supported 25 item-scale SAAS-R measures from five latent factors. Therefore, the advantages of multidimensional Rasch analysis are demonstrated in this study.</p>


2020 ◽  
pp. 135910532090476
Author(s):  
Natalie Papini ◽  
Minsoo Kang ◽  
Seungho Ryu ◽  
Emily Griese ◽  
Taylor Wingert ◽  
...  

Rasch modeling was used to examine the 25-item Connor-Davidson Resilience Scale within adults ( n = 410) in a weight management program. Rasch analysis assessed model-data fit, item difficulty and person’s resilience level, an item-person map to evaluate relative distribution items and persons, and rating scale function. Four misfit items were identified and removed. Item difficulty ranged from 1.25 to 1.19 logits (higher logit values indicate more difficult items). Persons’ resilience level had wide distribution (resilience = 2.27 ± 1.56 logits). Item difficulty levels did not adequately assess higher resilience levels. An improved inventory that measures a wider range of resilient behaviors would improve measurement quality.


2009 ◽  
Vol 6 (1) ◽  
pp. 205-205 ◽  
Author(s):  
K EVANS ◽  
K ANDERSON ◽  
B BOROWSKY ◽  
K DUFF ◽  
J GIULIANO ◽  
...  

2017 ◽  
Vol 38 (2) ◽  
pp. 68-76 ◽  
Author(s):  
Kristina Luhr ◽  
Ann Catrine Eldh ◽  
Ulrica Nilsson ◽  
Marie Holmefur

The Patient Preferences for Patient Participation tool (The 4Ps) was developed to aid clinical dialogue and to help patients to 1) depict, 2) prioritise, and 3) evaluate patient participation with 12 pre-set items reiterated in the three sections. An earlier qualitative evaluation of The 4Ps showed promising results. The present study is a psychometric evaluation of The 4Ps in patients with chronic heart or lung disease ( n = 108) in primary and outpatient care. Internal scale validity was evaluated using Rasch analysis, and two weeks test–retest reliability of the three sections using kappa/weighted kappa and a prevalence- and bias-adjusted kappa. The 4Ps tool was found to be reasonably valid with a varied reliability. Proposed amendments are rephrasing of two items, and modifications of the rating scale in Section 2. The 4Ps is suggested for use to increase general knowledge of patient participation, but further studies are needed with regards to its implementation.


2013 ◽  
Vol 93 (1) ◽  
pp. 60-68 ◽  
Author(s):  
David Walton ◽  
James M. Elliott

Background Despite increasing clinical and research use of the 11-item version of the Tampa Scale for Kinesiophobia (TSK-11) in people with neck pain, little is known about its measurement properties in this population. Objective The purpose of this study was to rigorously evaluate the measurement properties of the TSK-11 when used in people with mechanical neck pain. Design This study was a secondary analysis of 2 independent databases (N=235) of people with mechanical neck pain of primarily traumatic origin. Methods The TSK-11 was subjected to Rasch analysis and subsequent evaluation of concurrent associations with the Neck Disability Index and a numeric rating scale for pain intensity. Results The TSK-11 conformed well to the Rasch model for interval-level measurement, but less so for acute or nontraumatic etiologies. A transformation matrix suggested that small changes at the extremes of the scale are more meaningful than in the middle. Cross-sectional convergent validity testing suggested relationships of expected magnitude and direction compared with pain intensity and neck-related disability. The use of the linearly transformed TSK-11 led to potentially important differences in distribution of data compared with use of the raw scores. Limitations The sample size was slightly smaller than desired for Rasch analysis. The 2 databases were similar in terms of symptom duration, but differed in pain intensity and age. Conclusions The TSK-11 can be considered an interval-level measure when used in people with neck pain. It provides potentially important information regarding the nature of neck-related disability. Clinically important difference may not be consistent across the range of the scale.


Sign in / Sign up

Export Citation Format

Share Document