item response model
Recently Published Documents


TOTAL DOCUMENTS

188
(FIVE YEARS 45)

H-INDEX

23
(FIVE YEARS 2)

2021 ◽  
Vol 3 (4) ◽  
Author(s):  
Daniel Christen ◽  
Clare Killikelly ◽  
Andreas Maercker ◽  
Mareike Augsburger

Background In the 11th revision of the International Classification of Diseases (ICD-11) posttraumatic stress disorder (PTSD) and the complex variant (CPTSD) were newly conceptualised. The International Trauma Questionnaire (ITQ) was developed as a brief self-report measure to screen for both disorders. The English original version has been rigorously tested and presents convincing psychometric properties. The aim of the current study was to validate the German version by means of item response theory (IRT). Method This is a secondary analysis of a representative, trauma-exposed adult sample from the German general population (N = 500). 1- and 2-parameter logistic IRT models (i.e. examination on an item level), diagnostic rates and confirmatory factor analyses were calculated. Results All items showed good model fit and acceptable to good performance aligning with the items of the English original except for item C1 (Long time to calm down) which had a high endorsement rate and a low discriminatory power yielding low information gain. CPTSD diagnostic rate of 3.2% was lower than in comparable literature. Confirmatory factor analysis deemed the six first-order, two second-order factors model superior. Conclusion Measurement and factorial validity of the German version of the ITQ was confirmed. The German translation matches the English original in most psychometric properties and can thus be used for research and clinical practice.


2021 ◽  
Vol 10 (6) ◽  
pp. 103
Author(s):  
Sudarat Phaniew ◽  
Putcharee Junpeng ◽  
Keow Ngang Tang

This study intends to design and verify the quality of a model that measures mathematical proficiency and aims to set the standards in measuring levels of proficiency in the subjects of measurement and geometry. Construct modeling was employed to design a mathematical proficiency measurement model which consists of the mathematical process and the dimensions of a conceptual structure. A total of 517 Secondary Year 1 students were selected from the big data to participate as test-takers. Design-based research encompassing four phases was used to verify the quality of the mathematical proficiency measurement model. A Multidimensional Random Coefficient Multinomial Logit model was used to examine the standards-setting of the mathematical proficiency measurement model. The results indicated that the two dimensions of mathematical proficiency can be further divided into five levels, from non-response/irrelevance to strategic/extended thinking and extended abstract structure for mathematical process and conceptual structural dimensions, respectively. The assessment tool covers 18 items with 15 multiple-choice items and three subjective items in measurement and geometry. Moreover, the results also demonstrated that the validity evidence associated with the internal structure of the multidimensional model is fit. Besides, reliability evidence, as well as item fit, is compliance with the quality of the mathematical proficiency measurement model as illustrated in analysis of the standard error of measurement and infit and outfit of the items. Finally, the researchers managed to set standards for the mathematical proficiency measurement model based on the assessment criterion results from the Wright Map. In conclusion, the standards-setting of the mathematical proficiency measurement model provides substantial information, particularly for measuring those students who are above the lowest level of mathematical proficiency because the error for estimating proficiency was low.


SAGE Open ◽  
2021 ◽  
Vol 11 (4) ◽  
pp. 215824402110525
Author(s):  
Chanjin Zheng ◽  
Shaoyang Guo ◽  
Justin L Kern

There is a rekindled interest in the four-parameter logistic item response model (4PLM) after three decades of neglect among the psychometrics community. Recent breakthroughs in item calibration include the Gibbs sampler specially made for 4PLM and the Bayes modal estimation (BME) method as implemented in the R package mirt. Unfortunately, the MCMC is often time-consuming, while the BME method suffers from instability due to the prior settings. This paper proposes an alternative BME method, the Bayesian Expectation-Maximization-Maximization-Maximization (BE3M) method, which is developed from by combining an augmented variable formulation of the 4PLM and a mixture model conceptualization of the 3PLM. The simulation shows that the BE3M can produce estimates as accurately as the Gibbs sampling method and as fast as the EM algorithm. A real data example is also provided.


Foundations ◽  
2021 ◽  
Vol 1 (1) ◽  
pp. 116-144
Author(s):  
Alexander Robitzsch

This article investigates the comparison of two groups based on the two-parameter logistic item response model. It is assumed that there is random differential item functioning in item difficulties and item discriminations. The group difference is estimated using separate calibration with subsequent linking, as well as concurrent calibration. The following linking methods are compared: mean-mean linking, log-mean-mean linking, invariance alignment, Haberman linking, asymmetric and symmetric Haebara linking, different recalibration linking methods, anchored item parameters, and concurrent calibration. It is analytically shown that log-mean-mean linking and mean-mean linking provide consistent estimates if random DIF effects have zero means. The performance of the linking methods was evaluated through a simulation study. It turned out that (log-)mean-mean and Haberman linking performed best, followed by symmetric Haebara linking and a newly proposed recalibration linking method. Interestingly, linking methods frequently found in applications (i.e., asymmetric Haebara linking, recalibration linking used in a variant in current large-scale assessment studies, anchored item parameters, concurrent calibration) perform worse in the presence of random differential item functioning. In line with the previous literature, differences between linking methods turned out be negligible in the absence of random differential item functioning. The different linking methods were also applied in an empirical example that performed a linking of PISA 2006 to PISA 2009 for Austrian students. This application showed that estimated trends in the means and standard deviations depended on the chosen linking method and the employed item response model.


Psychometrika ◽  
2021 ◽  
Author(s):  
José H. Lozano ◽  
Javier Revuelta

AbstractThe present paper introduces a new explanatory item response model to account for the learning that takes place during a psychometric test due to the repeated use of the operations involved in the items. The proposed model is an extension of the operation-specific learning model (Fischer and Formann in Appl Psychol Meas 6:397–416, 1982; Scheiblechner in Z für Exp Angew Psychol 19:476–506, 1972; Spada in Spada and Kempf (eds.) Structural models of thinking and learning, Huber, Bern, Germany, pp 227–262, 1977). The paper discusses special cases of the model, which, together with the general formulation, differ in the type of response in which the model states that learning occurs: (1) correct and incorrect responses equally (non-contingent learning); (2) correct responses only (contingent learning); and (3) correct and incorrect responses to a different extent (differential contingent learning). A Bayesian framework is adopted for model estimation and evaluation. A simulation study is conducted to examine the performance of the estimation and evaluation methods in recovering the true parameters and selecting the true model. Finally, an empirical study is presented to illustrate the applicability of the model to detect learning effects using real data.


Psych ◽  
2021 ◽  
Vol 3 (3) ◽  
pp. 447-478
Author(s):  
Leah Feuerstahler

The filtered monotonic polynomial (FMP) model is a semi-parametric item response model that allows flexible response function shapes but also includes traditional item response models as special cases. The flexmet package for R facilitates the routine use of the FMP model in real data analysis and simulation studies. This tutorial provides several code examples illustrating how the flexmet package may be used to simulate FMP model parameters and data (both for dichotomous and polytomously scored items), estimate FMP model parameters, transform traditional item response models to different metrics, and more. This tutorial serves as both an introduction to the unique features of the FMP model and as a practical guide to its implementation in R via the flexmet package.


Author(s):  
Dan Cloney ◽  
Kellie Picker

Children develop rapidly in their early years. A crucial component of this development is a child’s ability to learn and use language. Even before they enter formal education, children have learned much about oral language and literacy through meaningful interactions with others, and from their life experiences. Children, however, do not develop at the same pace – some children arrive in early childhood education and care (ECEC) programs more advanced while others require additional support. Recent reviews of the assessment tools available to ECEC educators show a lack of good quality measurement and a reliance on checklist style inventories or narrative approaches. This paper presents a new measure of oral language and pre-literacy specifically designed to be accurate enough to reliably measure an individual child’s growth. Results from a combined calibration of children’s responses using a many-facets item response model show the measure to be reliable, valid and sensitive enough to measure growth within children and between groups of children over time. Implications for future assessment development and for educators’ practice are discussed, including how such measures can provide insight into what children know, understand, and can do (Reynolds, 2020) and what educators can do to support future learning experiences targeted at children’s specific language and literacy needs.


Mathematics ◽  
2021 ◽  
Vol 9 (13) ◽  
pp. 1465
Author(s):  
Alexander Robitzsch

This article shows that the recently proposed latent D-scoring model of Dimitrov is statistically equivalent to the two-parameter logistic item response model. An analytical derivation and a numerical illustration are employed for demonstrating this finding. Hence, estimation techniques for the two-parameter logistic model can be used for estimating the latent D-scoring model. In an empirical example using PISA data, differences of country ranks are investigated when using different metrics for the latent trait. In the example, the choice of the latent trait metric matters for the ranking of countries. Finally, it is argued that an item response model with bounded latent trait values like the latent D-scoring model might have advantages for reporting results in terms of interpretation.


Author(s):  
Alexander Robitzsch

This article shows that the recently proposed latent D-scoring model of Dimitrov is statistically equivalent to the two-parameter logistic item response model. An analytical derivation and a numerical illustration are employed for demonstrating this finding. Hence, estimation techniques for the two-parameter logistic model can be used for estimating the latent D-scoring model. In an empirical example using PISA data, differences of country ranks are investigated when using different metrics for the latent trait. In the example, the choice of the latent trait metric matters for the ranking of countries. Finally, it is argued that an item response model with bounded latent trait values like the latent D-scoring model might have advantages for reporting results in terms of interpretation.


Sign in / Sign up

Export Citation Format

Share Document