scholarly journals Polytomous versus Dichotomous Scoring on Multiple-Choice Examinations: Development of a Rubric for Rating Partial Credit

2013 ◽  
Vol 90 (10) ◽  
pp. 1310-1315 ◽  
Author(s):  
Megan L. Grunert ◽  
Jeffrey R. Raker ◽  
Kristen L. Murphy ◽  
Thomas A. Holme
Author(s):  
Michael Williams ◽  
Eileen Wood ◽  
Fatma Arslantas ◽  
Steve MacNeil

Multiple-choice testing with dichotomous scoring is one of the most common assessment methods utilized in undergraduate education. Determining students’ perceptions toward different types of multiple-choice testing formats is important for effective assessment. The present study compared two alternative multiple-choice testing formats used in a second-year required chemistry course: (1) The Immediate Feedback Assessment Technique (IFAT®) and (2) Personal Point Allocation (PPA). Both testing methods allow for partial credit but only the IFAT® provides immediate feedback on students’ responses. Both survey and interview data indicated that, overall, most students preferred IFAT® to the PPA testing method. These positive ratings were related to potential increase in reward, ease of use, and confidence. IFAT® was also perceived to be less stress producing, and anxiety provoking than PPA. Interview data supported these findings but also indicated individual differences in preference for each of these two methods. Additionally, students’ feedback on strategies used for either testing method and suggestions on how to improve the methods are discussed.


2013 ◽  
Vol 48 (1) ◽  
pp. 146-147 ◽  
Author(s):  
Yuan-chao Bo ◽  
Charles Lewis ◽  
David V. Budescu

2002 ◽  
Vol 90 (1) ◽  
pp. 226-226 ◽  
Author(s):  
Michael L. Epstein ◽  
Gary M. Brosvic

A multiple-choice testing system that provides immediate affirming or corrective feedback and permits allocation of partial credit for proximate knowledge is suggested as an alternative to essay examinations.


2021 ◽  
Vol 10 (2) ◽  
pp. 825-840
Author(s):  
Lukman Abdul ◽  
Syukrul Hamdi* ◽  
Masrid Pikoli ◽  
Romario Abdullah ◽  
Citra Panigoro

2017 ◽  
Vol 77 (1) ◽  
pp. 82-103 ◽  
Author(s):  
Georgios Sideridis ◽  
Ioannis Tsaousis ◽  
Khaleel Al Harbi

The purpose of the present article was to illustrate, using an example from a national assessment, the value from analyzing the behavior of distractors in measures that engage the multiple-choice format. A secondary purpose of the present article was to illustrate four remedial actions that can potentially improve the measurement of the construct(s) under study. Participants were 2,248 individuals who took a national examination of chemistry. The behavior of the distractors was analyzed by modeling their behavior within the Rasch model. Potentially informative distractors were (a) further modeled using the partial credit model, (b) split onto separate items and retested for model fit and parsimony, (c) combined to form a “super” item or testlet, and (d) reexamined after deleting low-ability individuals who likely guessed on those informative, albeit erroneous, distractors. Results indicated that all but the item split strategies were associated with better model fit compared with the original model. The best fitted model, however, involved modeling and crediting informative distractors via the partial credit model or eliminating the responses of low-ability individuals who likely guessed on informative distractors. The implications, advantages, and disadvantages of modeling informative distractors for measurement purposes are discussed.


2011 ◽  
Vol 72 (3) ◽  
pp. 493-509 ◽  
Author(s):  
Hong Jiao ◽  
Junhui Liu ◽  
Kathleen Haynie ◽  
Ada Woo ◽  
Jerry Gorham

This study explored the impact of partial credit scoring of one type of innovative items (multiple-response items) in a computerized adaptive version of a large-scale licensure pretest and operational test settings. The impacts of partial credit scoring on the estimation of the ability parameters and classification decisions in operational test settings were explored in one real data analysis and two simulation studies when two different polytomous scoring algorithms, automated polytomous scoring and rater-generated polytomous scoring, were applied. For the real data analyses, the ability estimates from dichotomous and polytomous scoring were highly correlated; the classification consistency between different scoring algorithms was nearly perfect. Information distribution changed slightly in the operational item bank. In the two simulation studies comparing each polytomous scoring with dichotomous scoring, the ability estimates resulting from polytomous scoring had slightly higher measurement precision than those resulting from dichotomous scoring. The practical impact related to classification decision was minor because of the extremely small number of items that could be scored polytomously in this current study.


2017 ◽  
Vol 33 (5) ◽  
pp. 336-344 ◽  
Author(s):  
Birk Diedenhofen ◽  
Jochen Musch

Abstract. Standard dichotomous scoring of multiple-choice test items grants no partial credit for partial knowledge. Empirical option weighting is an alternative, polychotomous scoring method that uses the point-biserial correlation between option choices and total score as a weight for each answer alternative. Extant studies demonstrate that the method increases reliability of multiple-choice tests in comparison to conventional scoring. Most previous studies employed a correlational validation approach, however, and provided mixed findings with regard to the validity of empirical option weighting. The present study is the first investigation using an experimental approach to determine the reliability and validity of empirical option weighting. To obtain an external validation criterion, we experimentally induced various degrees of knowledge in a domain of which participants had no knowledge. We found that in comparison to dichotomous scoring, empirical option weighting increased both reliability and validity of a multiple-choice knowledge test employing distractors that were appealing to test takers with different levels of knowledge. A potential application of the present results is the computation and publication of empirical option weights for existing multiple-choice knowledge tests that have previously been scored dichotomously.


1995 ◽  
Vol 77 (3) ◽  
pp. 760-762
Author(s):  
Kenneth S. Shultz

Little research has been conducted on the use of linear polychotomous scoring of multiple-choice test items. Therefore, several tests were analyzed using both dichotomous and polychotomous scoring of test items to assess how the alpha reliabilities of the tests change based on the type of scoring used. In each case, the alpha reliabilities of the tests increased, with the same or fewer number of items in each test, when polychotomous (vs dichotomous) scoring of multiple-choice test items was used.


Sign in / Sign up

Export Citation Format

Share Document