TOEFL Scores: What Changed and What Stayed—A Lateral Study of 2017~2020 TOEFL iBT Test and Score Data Summary

2021 ◽  
Vol 03 (03) ◽  
pp. 119-127
Author(s):  
宇西 姚
Keyword(s):  
2021 ◽  
pp. 026553222110033
Author(s):  
Carsten Roever ◽  
Naoki Ikeda

The overarching aim of the study is to explore the extent to which test takers’ performances on monologic speaking tasks provide information about their interactional competence. This is an important concern from a test use perspective, as stakeholders tend to consider test scores as providing comprehensive information about all aspects of L2 competence. One hundred and fifty test takers completed a TOEFL iBT speaking section consisting of six monologic tasks, measuring speaking proficiency, followed by a test of interactional competence with three monologues and three dialogues, measuring pragmalinguistic skills, the ability to recipient design extended discourse, and interactional management skills. Quantitative analyses showed a medium to high correlation between TOEFL iBT speaking scores and interactional scores of r = .76, though with a much lower correlation of r = .57 for the subsample most similar to a typical TOEFL population. There was a large amount of variation in interactional scores for test takers at the same TOEFL iBT speaking score level, and qualitative analyses demonstrated that test takers’ ability to recipient design their talk and format social actions appropriate to social roles and relationships was not well captured by speaking scores. We suggest potential improvements.


2010 ◽  
Vol 27 (3) ◽  
pp. 335-353 ◽  
Author(s):  
Sara Cushing Weigle

Automated scoring has the potential to dramatically reduce the time and costs associated with the assessment of complex skills such as writing, but its use must be validated against a variety of criteria for it to be accepted by test users and stakeholders. This study approaches validity by comparing human and automated scores on responses to TOEFL® iBT Independent writing tasks with several non-test indicators of writing ability: student self-assessment, instructor assessment, and independent ratings of non-test writing samples. Automated scores were produced using e-rater ®, developed by Educational Testing Service (ETS). Correlations between both human and e-rater scores and non-test indicators were moderate but consistent, providing criterion-related validity evidence for the use of e-rater along with human scores. The implications of the findings for the validity of automated scores are discussed.


2021 ◽  
Vol 50 (1) ◽  
Author(s):  
Scott Roy Douglas ◽  
Michael Landry

Because of the large number of post-secondary English for academic purposes (EAP) programs and the varying ways they are structured, it can be difficult to identify how a particular program fits within the overall landscape of university education.  To identify general trends across Canada, the webpages for 74 EAP programs at 50 public English-medium universities were examined for key information related to each program.  Data analysis included descriptive statistics as well as graphical representation.  The results pointed to typical EAP programs that are independent units that offer non-credit courses with some credit options, have international tuition fees around $9,000 per semester, provide approximately 22 hours of instruction per week, and generally require IELTS scores over 5.0 or TOEFL iBT scores over 59 for entry.  These results provide an avenue of comparison and indicate the need for future research to better understand how EAP programming is conceptualized in the Canadian context.


2019 ◽  
Vol 16 (2) ◽  
pp. 253-270 ◽  
Author(s):  
Annie Brown ◽  
Ana Maria Ducasse
Keyword(s):  

2019 ◽  
Vol 54 (1) ◽  
pp. 251-265
Author(s):  
Lorena Llosa ◽  
Scott E. Grapin ◽  
Eric Friginal ◽  
Sara T. Cushing ◽  
Margaret E. Malone

Sign in / Sign up

Export Citation Format

Share Document