Fitting a Two-Parameter Logistic Item Response Model to Clarify the Psychometric Properties of the Drug Use Screening Inventory for Adolescent Alcohol and Drug Abusers

1994 ◽  
Vol 18 (6) ◽  
pp. 1335-1341 ◽  
Author(s):  
Levent Kirisci ◽  
Ralph E. Tarter ◽  
Tse-Chi Hsu
2019 ◽  
Vol 80 (3) ◽  
pp. 604-612
Author(s):  
Tenko Raykov ◽  
George A. Marcoulides

This note raises caution that a finding of a marked pseudo-guessing parameter for an item within a three-parameter item response model could be spurious in a population with substantial unobserved heterogeneity. A numerical example is presented wherein each of two classes the two-parameter logistic model is used to generate the data on a multi-item measuring instrument, while the three-parameter logistic model is found to be associated with a considerable pseudo-guessing parameter estimate on an item. The implications of the reported results for empirical educational research are subsequently discussed.


Author(s):  
Alexander Robitzsch

This article shows that the recently proposed latent D-scoring model of Dimitrov is statistically equivalent to the two-parameter logistic item response model. An analytical derivation and a numerical illustration are employed for demonstrating this finding. Hence, estimation techniques for the two-parameter logistic model can be used for estimating the latent D-scoring model. In an empirical example using PISA data, differences of country ranks are investigated when using different metrics for the latent trait. In the example, the choice of the latent trait metric matters for the ranking of countries. Finally, it is argued that an item response model with bounded latent trait values like the latent D-scoring model might have advantages for reporting results in terms of interpretation.


Foundations ◽  
2021 ◽  
Vol 1 (1) ◽  
pp. 116-144
Author(s):  
Alexander Robitzsch

This article investigates the comparison of two groups based on the two-parameter logistic item response model. It is assumed that there is random differential item functioning in item difficulties and item discriminations. The group difference is estimated using separate calibration with subsequent linking, as well as concurrent calibration. The following linking methods are compared: mean-mean linking, log-mean-mean linking, invariance alignment, Haberman linking, asymmetric and symmetric Haebara linking, different recalibration linking methods, anchored item parameters, and concurrent calibration. It is analytically shown that log-mean-mean linking and mean-mean linking provide consistent estimates if random DIF effects have zero means. The performance of the linking methods was evaluated through a simulation study. It turned out that (log-)mean-mean and Haberman linking performed best, followed by symmetric Haebara linking and a newly proposed recalibration linking method. Interestingly, linking methods frequently found in applications (i.e., asymmetric Haebara linking, recalibration linking used in a variant in current large-scale assessment studies, anchored item parameters, concurrent calibration) perform worse in the presence of random differential item functioning. In line with the previous literature, differences between linking methods turned out be negligible in the absence of random differential item functioning. The different linking methods were also applied in an empirical example that performed a linking of PISA 2006 to PISA 2009 for Austrian students. This application showed that estimated trends in the means and standard deviations depended on the chosen linking method and the employed item response model.


Mathematics ◽  
2021 ◽  
Vol 9 (13) ◽  
pp. 1465
Author(s):  
Alexander Robitzsch

This article shows that the recently proposed latent D-scoring model of Dimitrov is statistically equivalent to the two-parameter logistic item response model. An analytical derivation and a numerical illustration are employed for demonstrating this finding. Hence, estimation techniques for the two-parameter logistic model can be used for estimating the latent D-scoring model. In an empirical example using PISA data, differences of country ranks are investigated when using different metrics for the latent trait. In the example, the choice of the latent trait metric matters for the ranking of countries. Finally, it is argued that an item response model with bounded latent trait values like the latent D-scoring model might have advantages for reporting results in terms of interpretation.


2014 ◽  
Vol 28 (1) ◽  
pp. 1-23 ◽  
Author(s):  
Jorge Luis Bazán ◽  
Márcia D. Branco ◽  
Heleno Bolfarine

2011 ◽  
Vol 6 (3) ◽  
pp. 354-398 ◽  
Author(s):  
Katharine O. Strunk

Increased spending and decreased student performance have been attributed in part to teachers' unions and to the collective bargaining agreements (CBAs) they negotiate with school boards. However, only recently have researchers begun to examine impacts of specific aspects of CBAs on student and district outcomes. This article uses a unique measure of contract restrictiveness generated through the use of a partial independence item response model to examine the relationships between CBA strength and district spending on multiple areas and district-level student performance in California. I find that districts with more restrictive contracts have higher spending overall, but that this spending appears not to be driven by greater compensation for teachers but by greater expenditures on administrators' compensation and instruction-related spending. Although districts with stronger CBAs spend more overall and on these categories, they spend less on books and supplies and on school board–related expenditures. In addition, I find that contract restrictiveness is associated with lower average student performance, although not with decreased achievement growth.


Sign in / Sign up

Export Citation Format

Share Document