scholarly journals Evaluating Item Difficulty Patterns for Assessing Student Misconceptions in Science Across Physics, Chemistry, and Biology Concepts

2021 ◽  
Author(s):  
Soeharto Soeharto ◽  
Benő Csapó
Biosfer ◽  
2021 ◽  
Vol 14 (1) ◽  
pp. 99-119
Author(s):  
Feni Andriani ◽  
Meti Indrowati ◽  
Bowo Sugiharto

The purpose of this study was to analyze the feasibility of the items of the four-tier multiple-choice test immune system instrument that had been developed. The development of the instrument using the Treagust (1988) model, namely defining content, collecting student misconceptions information, and developing a diagnostic test. A total of 25 items have been developed. The results of the instrument development were tested on 142 students of grade XI from several high schools in Surakarta who were selected by simple random sampling. The data analysis technique was performed using Rasch analysis in the Winstep application. The results of the construct validity test showed items number 5, 7, and 9 did not fit the validity standards. The reliability test shows that the value of Cronbach Alpha reliability is bad (n = 0.51), the value of the reliability item is special (no = 0.97), the value of person reliability is sufficient (n = 0.68), the value of person separation is weak (n = 1.44), and the item separation value is special (n = 5.38). The person discrimination test showed student 056P31 has the highest ability and student 098P51 has the lowest ability. The item discrimination test shows item number 1 is the best item and the bad item is number 14. The item difficulty analysis showed less proportionality because there were too many items in the easy and difficult categories. An expansion of the sample is needed to see a more comprehensive and diverse range of responses to instruments.


2020 ◽  
Vol 36 (4) ◽  
pp. 554-562
Author(s):  
Alica Thissen ◽  
Frank M. Spinath ◽  
Nicolas Becker

Abstract. The cube construction task represents a novel format in the assessment of spatial ability through mental cube rotation tasks. Instead of selecting the correct answer from several response options, respondents construct their own response in a computerized test environment, leading to a higher demand for spatial ability. In the present study with a sample of 146 German high-school students, we tested an approach to manipulate the item difficulties in order to create items with a greater difficulty range. Furthermore, we compared the cube task in a distractor-free and a distractor-based version while the item stems were held identical. The average item difficulty of the distractor-free format was significantly higher than in the distractor-based format ( M = 0.27 vs. M = 0.46) and the distractor-free format showed a broader range of item difficulties (.02 ≤  pi ≤ .95 vs. .37 ≤  pi ≤ .63). The analyses of the test results also showed that the distractor-free format had a significantly higher correlation with a broad intelligence test ( r = .57 vs. r = .17). Reasons for the higher convergent validity of the distractor-free format (prevention of response elimination strategies and the broader range of item difficulties) and further research possibilities are discussed.


2012 ◽  
Author(s):  
Victoria Blanshteyn ◽  
Charles A. Scherbaum
Keyword(s):  

2010 ◽  
Vol 65 (4) ◽  
pp. 257-282
Author(s):  
전유아 ◽  
신택수

Sign in / Sign up

Export Citation Format

Share Document