ITEM BANKING FOR AN ADAPTIVE MEASUREMENT OF NEUROTICISM

2020 ◽  
Vol 13 (2) ◽  
pp. 459-485
Author(s):  
Facundo Juan Pablo Abal ◽  
Gabriela Susana Lozzia ◽  
Sofía Esmeralda Auné ◽  
Horacio Félix Attorresi

The psychometric properties of a bank of 36 items are presented measuring Neuroticism based on the Five-Factor Model. These items pertain to the facets that were identified by the work of McCrae and Costa. The sample was comprised of 1133 adult subjects that reside in the Buenos Aires Metropolitan Area in Argentina. Women accounted for 52.1% of those subjects with an average age of 29.5 years (SD = 11.32). In order to get the items calibrated according to Item Response Theory (Graded Response Model), acquire the bank’s information functions and assess the estimated associations with other instruments, 70% of the cases were randomly selected. An adaptive administration simulation was made with the remaining 30% so as to test two stopping rules: a) using 18 items and b) standard error of ≤ 0.25. Correlations greater than .95 were found between the estimated bank scores and the two adaptive versions. The advantages of using the adaptive Neuroticism measurement over other well-renowned instruments that use conventional large formats, as well as abbreviated ones, are discussed.

2018 ◽  
Vol 79 (3) ◽  
pp. 545-557 ◽  
Author(s):  
Dimiter M. Dimitrov ◽  
Yong Luo

An approach to scoring tests with binary items, referred to as D-scoring method, was previously developed as a classical analog to basic models in item response theory (IRT) for binary items. As some tests include polytomous items, this study offers an approach to D-scoring of such items and parallels the results with those obtained under the graded response model (GRM) for ordered polytomous items in the framework of IRT. The proposed design of using D-scoring with “virtual” binary items generated from polytomous items provides (a) ability scores that are consistent with their GRM counterparts and (b) item category response functions analogous to those obtained under the GRM. This approach provides a unified framework for D-scoring and psychometric analysis of tests with binary and/or polytomous items that can be efficient in different scenarios of educational and psychological assessment.


2018 ◽  
Vol 43 (4) ◽  
pp. 290-302 ◽  
Author(s):  
Dirk Lubbe ◽  
Christof Schuster

Questionnaires with uniform-ordered categorical response formats are widely applied in psychology. Muraki proposed a modified graded response model accounting for the items’ uniform response formats by assuming identical threshold parameters defining the category boundaries for all items. What is not well known is that there is a set of closely related models, which similarly assume identical thresholds. The present article gives a framework illustrating the differences between these models and their utility for understanding questionnaire responses in detail. The models are explained as constrained cases of a one-dimensional factor model for ordered categorical data. Furthermore, the authors show that the models can be written and fitted as structural equation models, which allows for a very flexible and general purpose use. Instructions on implementing the models in Mplus and SAS PROC NLMIXED are given.


2016 ◽  
Vol 59 (2) ◽  
pp. 373-383 ◽  
Author(s):  
J. Mirjam Boeschen Hospers ◽  
Niels Smits ◽  
Cas Smits ◽  
Mariska Stam ◽  
Caroline B. Terwee ◽  
...  

Purpose We reevaluated the psychometric properties of the Amsterdam Inventory for Auditory Disability and Handicap (AIADH; Kramer, Kapteyn, Festen, & Tobi, 1995) using item response theory. Item response theory describes item functioning along an ability continuum. Method Cross-sectional data from 2,352 adults with and without hearing impairment, ages 18–70 years, were analyzed. They completed the AIADH in the web-based prospective cohort study “Netherlands Longitudinal Study on Hearing.” A graded response model was fitted to the AIADH data. Category response curves, item information curves, and the standard error as a function of self-reported hearing ability were plotted. Results The graded response model showed a good fit. Item information curves were most reliable for adults who reported having hearing disability and less reliable for adults with normal hearing. The standard error plot showed that self-reported hearing ability is most reliably measured for adults reporting mild up to moderate hearing disability. Conclusions This is one of the few item response theory studies on audiological self-reports. All AIADH items could be hierarchically placed on the self-reported hearing ability continuum, meaning they measure the same construct. This provides a promising basis for developing a clinically useful computerized adaptive test, where item selection adapts to the hearing ability of individuals, resulting in efficient assessment of hearing disability.


2020 ◽  
Vol 9 (11) ◽  
pp. 3754
Author(s):  
Yoshiaki Nomura ◽  
Toshiya Morozumi ◽  
Mitsuo Fukuda ◽  
Nobuhiro Hanada ◽  
Erika Kakuta ◽  
...  

Periodontal examination data have a complex structure. For epidemiological studies, mass screenings, and public health use, a simple index that represents the periodontal condition is necessary. Periodontal indices for partial examination of selected teeth have been developed. However, the selected teeth vary between indices, and a justification for the selection of examination teeth has not been presented. We applied a graded response model based on the item response theory to select optimal examination teeth and sites that represent periodontal conditions. Data were obtained from 254 patients who participated in a multicenter follow-up study. Baseline data were obtained from initial follow-up. Optimal examination sites were selected using item information calculated by graded response modeling. Twelve sites—maxillary 2nd premolar (palatal-medial), 1st premolar (palatal-distal), canine (palatal-medial), lateral incisor (palatal-central), central incisor (palatal-distal) and mandibular 1st premolar (lingual, medial)—were selected. Mean values for clinical attachment level, probing pocket depth, and bleeding on probing by full mouth examinations were used for objective variables. Measuring the clinical parameters of these sites can predict the results of full mouth examination. For calculating the periodontal index by partial oral examination, a justification for the selection of examination sites is essential. This study presents an evidence-based partial examination methodology and its modeling.


2021 ◽  
Vol 9 ◽  
pp. 205031212110122
Author(s):  
Samuel W Terman ◽  
James F Burke

Objectives: Accurately measuring disability is critical toward policy development, economic analyses, and determining individual-level effects of health interventions. Nationally representative population surveys such as the National Health and Nutrition Examination Survey provide key opportunities to measure disability constructs such as activity limitations. However, only very limited work has previously evaluated the item response properties of questions pertaining to limitations in National Health and Nutrition Examination Survey. Methods: This was a cross-sectional study. We included participants ⩾20 years old for the 2013–2018 National Health and Nutrition Examination Survey cycles. Activity limitations, and a smaller number of body function impairments or participation restrictions, were determined from interview questions. We performed item response theory models (a two-parameter logistic and a graded response model) to characterize discriminating information along the latent continuum of activity limitation. Results: We included 17,057 participants. Although each particular limitation was somewhat rare (maximally 13%), 7214 (38%) responded having at least one limitation. We found a high amount of discriminating information at 1–2 standard deviations above average limitation, though essentially zero discrimination below that range. Items had substantial overlap in the range at which they provided information distinguishing individuals. The ordinal graded response model including 20 limitations provided greater information than the dichotomous two-parameter logistic model, though further omitting items from the graded response model led to loss of information. Conclusion: National Health and Nutrition Examination Survey disability-related questions, mostly specifically activity limitations, provided a high degree of information distinguishing individuals with higher than average limitations on the latent continuum, but essentially zero resolution to distinguish individuals with low or average limitations. Future work may focus on developing items which better distinguish individuals at the “lower” end of the limitation spectrum.


2017 ◽  
Vol 78 (3) ◽  
pp. 384-408 ◽  
Author(s):  
Yong Luo ◽  
Hong Jiao

Stan is a new Bayesian statistical software program that implements the powerful and efficient Hamiltonian Monte Carlo (HMC) algorithm. To date there is not a source that systematically provides Stan code for various item response theory (IRT) models. This article provides Stan code for three representative IRT models, including the three-parameter logistic IRT model, the graded response model, and the nominal response model. We demonstrate how IRT model comparison can be conducted with Stan and how the provided Stan code for simple IRT models can be easily extended to their multidimensional and multilevel cases.


2017 ◽  
Vol 14 (1) ◽  
pp. 113-117 ◽  
Author(s):  
Daniel S.J. Costa ◽  
Ali Asghari ◽  
Michael K. Nicholas

AbstractBackground and aimsThe Pain Self-Efficacy Questionnaire (PSEQ) is a 10-item instrument designed to assess the extent to which a person in pain believes s/he is able to accomplish various activities despite their pain. There is strong evidence for the validity and reliability of both the full-length PSEQ and a 2-item version. The purpose of this study is to further examine the properties of the PSEQ using an item response theory (IRT) approach.MethodsWe used the two-parameter graded response model to examine the category probability curves, and location and discrimination parameters of the 10 PSEQ items. In item response theory, responses to a set of items are assumed to be probabilistically determined by a latent (unobserved) variable. In the graded-response model specifically, item response threshold (the value of the latent variable for which adjacent response categories are equally likely) and discrimination parameters are estimated for each item. Participants were 1511 mixed, chronic pain patients attending for initial assessment at a tertiary pain management centre.ResultsAll items except item 7 (‘I can cope with my pain without medication’) performed well in IRT analysis, and the category probability curves suggested that participants used the 7-point response scale consistently. Items 6 (‘I can still do many of the things I enjoy doing, such as hobbies or leisure activity, despite pain’), 8 (‘I can still accomplish most of my goals in life, despite the pain’) and 9 (‘I can live a normal lifestyle, despite the pain’) captured higher levels of the latent variable with greater precision.ConclusionsThe results from this IRT analysis add to the body of evidence based on classical test theory illustrating the strong psychometric properties of the PSEQ. Despite the relatively poor performance of Item 7, its clinical utility warrants its retention in the questionnaire.ImplicationsThe strong psychometric properties of the PSEQ support its use as an effective tool for assessing self-efficacy in people with pain.


2020 ◽  
Vol 18 (1) ◽  
Author(s):  
Fengling Yang ◽  
Fei Zhao ◽  
Yun Zheng ◽  
Gang Li

Abstract Background Early prelingual auditory development (EPLAD) is a fundamental and important process in the speech and language development of infants and toddlers. The Infant–Toddler Meaningful Auditory Integration Scale (ITMAIS) is a widely used measurement tool for EPLAD, however it has not yet undergone a comprehensive psychometric analysis. The aim of this research was to modify and verify the psychometric properties of ITMAIS using a combination of Item Response Theory (IRT) and Classical Test Theory (CTT). Methods Stage 1—1730 children were retrospectively recruited to enable the application of an IRT model, specifically the graded response model, to modify the ITMAIS. Stage 2—another 450 infants and toddlers with normal hearing or permanent hearing loss before auditory intervention were recruited to verify the psychometric properties of the modified ITMAIS (ITMAIS-m) using the CTT method. Results Using the metric of the graded response model, by removing item 2 from the ITMAIS, ITMAIS-m demonstrated discrimination parameters ranging from 3.947 to 5.431, difficulty parameters from − 1.146 to 1.150, item information distributed between 4.798 and 9.259 and a test information score of 48.061. None of the items showed differential item functioning. ITMAIS-m was further verified in Stage 2, showing Cronbach’s α of 0.919 and item-total correlations ranging from 0.693 to 0.851. There was good convergent validity of ITMAIS-m with other auditory outcome measure (r = 0.932) and pure tone average thresholds (r ranging from − 0.670 to − 0.909), as well as a high ability to discriminate between different hearing grades (Cohen d ranging from 0.41 to 5.83). Conclusions The ITMAIS-m is a reliable and valid tool for evaluating EPLAD in infants and toddlers, which can be efficiently and precisely applied in clinical practice. The combined use of IRT and CTT provides a powerful means to modify psychometrically robust scales aimed at childhood auditory outcome measurements.


Sign in / Sign up

Export Citation Format

Share Document