LinkIt: A Tool for Assessing and Providing Formative Feedback for Structural Knowledge

Author(s):  
Russell J. Branaghan ◽  
Roger W. Schvaneveldt ◽  
Jennifer L. Winner

It is challenging to assess the effectiveness of learning and training. Most evaluators rely on multiple choice, fill-in-the-blank, and essay questions. These are time-consuming, provide little formative value, and provide no visualization of the knowledge domain. LinkIt uses constrained concept mapping to elicit, score, represent visually, and provide immediate feedback on student knowledge networks by comparing them to expert networks.

2020 ◽  
Vol 3 (4) ◽  
pp. 276-287
Author(s):  
The Vinh Tran ◽  
Tran Kim Thanh ◽  
Tran Manh Tuong ◽  
Vu Anh Linh Duy

In Vietnam, since 2015, the Ministry of Education and Training of Vietnam has decided to abolish university entrance exams and advocates the use of high school graduation exam results of candidates for admission to go to universities. The 2015 and 2016 exam questions for the Math exam are the essay questions. From 2017 up to now, the Ministry of Education and Training of Vietnam has applied the form of multiple-choice exams for Mathematics in the high school graduation exam. There are many mixed opinions about the impact of this form of examination and admission on the quality of university students. In particular, the switch from the form of essay examination to multiple-choice exams led the entire Vietnam Mathematical Association at that time to send recommendations on continuing to maintain the form of essay examination for mathematics. The purposes of this article are analysis and evaluation the effects of relevant factors on the academic performance of advanced math students of university students, and offer solutions to optimize university entrance exam. The data set was provided by Training Management Department and Training Quality Control and Testing Laboratory of the University of Finance – Marketing. This dataset includes information about math high school graduation test scores, learning process scores (scores assessed by direct instructors), and advanced math course end test scores of 2834 students in courses from 2015 to 2019. Linear and non-linear regression machine learning models were used to solve the tasks given in this article. An analysis of the data was conducted to reveal the advantages and disadvantages of the change in university enrollment of the Vietnamese Ministry of Education and Training. Tools from the Python libraries have been supported and used effectively in the process of solving problems. Through building and surveying the model, there are suggestions and solutions to problems in enrollment and input quality assurance. Specifically, in the preparation of entrance exams, the entrance exam questions should not exceed 61-66 % of multiple choice questions.


2010 ◽  
Vol 34 (3) ◽  
pp. 452-458 ◽  
Author(s):  
William Rafaelo Schlinkert ◽  
Sandro Scarpelini ◽  
Antonio Pazin-Filho

BACKGROUND: E-learning techniques are spreading at great speed in medicine, raising concerns about the impact of adopting them. Websites especially designed to host courses are becoming more common. There is a lack of evidence that these systems could enhance student knowledge acquisition. GOAL: To evaluate the impact of using dedicated-website tools over cognition of medical students exposed to a first-aid course. METHODS: Prospective study of 184 medical students exposed to a twenty-hour first-aid course. We generated a dedicated-website with several sections (lectures, additional reading material, video and multiple choice exercises). We constructed variables expressing the student's access to each section. The evaluation was composed of fifty multiple-choice tests, based on clinical problems. We used multiple linear regression to adjust for potential confounders. RESULTS: There was no association of website intensity of exposure and the outcome - beta-coeficient 0.27 (95%CI - 0.454 - 1.004). These findings were not altered after adjustment for potential confounders - 0.165 (95%CI -0.628 - 0.960). CONCLUSION: A dedicated website with passive and active capabilities for aiding in person learning had not shown association with a better outcome.


2013 ◽  
Vol 77 (12) ◽  
pp. 1593-1609 ◽  
Author(s):  
C. Alex McMahan ◽  
R. Neal Pinckard ◽  
Thomas J. Prihoda ◽  
William D. Hendricson ◽  
Anne Cale Jones

2019 ◽  
Vol 16 (1) ◽  
pp. 59-73 ◽  
Author(s):  
Peter McKenna

PurposeThis paper aims to examine whether multiple choice questions (MCQs) can be answered correctly without knowing the answer and whether constructed response questions (CRQs) offer more reliable assessment.Design/methodology/approachThe paper presents a critical review of existing research on MCQs, then reports on an experimental study where two objective tests (using MCQs and CRQs) were set for an introductory undergraduate course. To maximise completion, tests were kept short; consequently, differences between individuals’ scores across both tests are examined rather than overall averages and pass rates.FindingsMost students who excelled in the MCQ test did not do so in the CRQ test. Students could do well without necessarily understanding the principles being tested.Research limitations/implicationsConclusions are limited by the small number of questions in each test and by delivery of the tests at different times. This meant that statistical average data would be too coarse to use, and that some students took one test but not the other. Conclusions concerning CRQs are limited to disciplines where numerical answers or short and constrained text answers are appropriate.Practical implicationsMCQs, while useful in formative assessment, are best avoided for summative assessments. Where appropriate, CRQs should be used instead.Social implicationsMCQs are commonplace as summative assessments in education and training. Increasing the use of CRQs in place of MCQs should increase the reliability of tests, including those administered in safety-critical areas.Originality/valueWhile others have recommended that MCQs should not be used (Hinchliffe 2014, Srivastavaet al., 2004) because they are vulnerable to guessing, this paper presents an experimental study designed to demonstrate whether this hypothesis is correct.


2002 ◽  
Vol 90 (1) ◽  
pp. 226-226 ◽  
Author(s):  
Michael L. Epstein ◽  
Gary M. Brosvic

A multiple-choice testing system that provides immediate affirming or corrective feedback and permits allocation of partial credit for proximate knowledge is suggested as an alternative to essay examinations.


Author(s):  
Michael Williams ◽  
Eileen Wood ◽  
Fatma Arslantas ◽  
Steve MacNeil

Multiple-choice testing with dichotomous scoring is one of the most common assessment methods utilized in undergraduate education. Determining students’ perceptions toward different types of multiple-choice testing formats is important for effective assessment. The present study compared two alternative multiple-choice testing formats used in a second-year required chemistry course: (1) The Immediate Feedback Assessment Technique (IFAT®) and (2) Personal Point Allocation (PPA). Both testing methods allow for partial credit but only the IFAT® provides immediate feedback on students’ responses. Both survey and interview data indicated that, overall, most students preferred IFAT® to the PPA testing method. These positive ratings were related to potential increase in reward, ease of use, and confidence. IFAT® was also perceived to be less stress producing, and anxiety provoking than PPA. Interview data supported these findings but also indicated individual differences in preference for each of these two methods. Additionally, students’ feedback on strategies used for either testing method and suggestions on how to improve the methods are discussed.


2020 ◽  
Vol 4 (2) ◽  
pp. 20
Author(s):  
María José García Tárrago

Existe gran número de publicaciones en relación con la fiabilidad de los test multi-respuesta para la evaluación del alumnado en la educación superior. Número de opciones por pregunta, sistemas de puntuación (marcado positivo o negativo), puntuación del conocimiento parcial o cantidad total de preguntas… La combinación de todos estos parámetros es una muestra de la variedad de configuraciones que pueden llegar a establecerse al diseñar un test. ¿Existe algún modelo o configuración óptima? Durante años, los investigadores en innovación educativa han intentado responder a esta cuestión haciendo uso del cálculo de probabilidades y distintas evaluaciones empíricas.En esta investigación se ha desarrollado un algoritmo basado en código Python con la finalidad de generar una serie de estudiantes hipotéticos con características y habilidades específicas (conocimiento real, nivel de cautela…). Un alto nivel de conocimientos implicaría una alta probabilidad de saber si una de las opciones de respuesta a una cuestión es cierta o no. Un exceso en el nivel de cautela de un alumno estaría relacionado con el nivel de probabilidad que lleva al alumno a arriesgarse a responder a una pregunta de la que no tiene por seguro su respuesta. Ello sería una medida de la capacidad de riesgo del alumno. El algoritmo lanza test a un número específico de alumnos hipotéticos analizando la desviación existente entre el conocimiento real (una característica intrínseca de cada alumno), y el conocimiento estimado por el test.Una vez desarrollado el algoritmo, se buscó validarlo con el uso de los distintos parámetros de entrada con la finalidad de observar la influencia que estos tenían en la puntuación final del test.AbstractThere are many literatures related with the reliability of true/false and multiple- choice tests and their application in higher education. Choices per question, positive or negative marking, rewards of partial knowledge or how long they should be… The combination of all these parameters shows the wide set of test setup that each examiner could design. Is there any optimized configuration? An extended educational research has tried to answer these questions using probability calculations and empirical evaluations.In this investigation, a novel algorithm was designed with Python code to generate hypothetical examinees with specific features (real knowledge, degree of over-cautiousness, fatigue limit…). High knowledge level implies high probability to know whether an answer choice was true or false in a multiple- choice question. Over-cautiousness was related with the probability to answer an unknown question or the risk capacity of the examinee. Finally, fatigue is directly related with the number of questions in the test. Going beyond its upper limit the knowledge level is reduced and the over-cautiousness is increased. The algorithm launched tests to the hypothetical examinees analysing the deviation between the real knowledge (a feature of the examinee), and the estimated knowledge.This algorithm was used to optimize the different parameters of a test (length of test, choices per question, scoring system…) to reduce the influence of fatigue and over-cautiousness on the final score. An empirical evaluation was performed comparing different test setups to verify and validate the algorithm.


Sign in / Sign up

Export Citation Format

Share Document