scholarly journals CoNLL 2014 Shared Task: Grammatical Error Correction with a Syntactic N-gram Language Model from a Big Corpora

Author(s):  
S. David Hdez. ◽  
Hiram Calvo
2020 ◽  
Vol 34 (10) ◽  
pp. 13859-13860
Author(s):  
Yiyuan Li ◽  
Antonios Anastasopoulos ◽  
Alan W. Black

Current grammatical error correction (GEC) models typically consider the task as sequence generation, which requires large amounts of annotated data and limit the applications in data-limited settings. We try to incorporate contextual information from pre-trained language model to leverage annotation and benefit multilingual scenarios. Results show strong potential of Bidirectional Encoder Representations from Transformers (BERT) in grammatical error correction task.


2018 ◽  
Vol 24 (6) ◽  
pp. 301-306
Author(s):  
Seung Woo Cho ◽  
Hong-seok Kwon ◽  
Hun-young Jung ◽  
Jong-Hyeok Lee

2019 ◽  
Author(s):  
Christopher Bryant ◽  
Mariano Felice ◽  
Øistein E. Andersen ◽  
Ted Briscoe

2014 ◽  
Vol 2 ◽  
pp. 419-434 ◽  
Author(s):  
Alla Rozovskaya ◽  
Dan Roth

This paper identifies and examines the key principles underlying building a state-of-the-art grammatical error correction system. We do this by analyzing the Illinois system that placed first among seventeen teams in the recent CoNLL-2013 shared task on grammatical error correction. The system focuses on five different types of errors common among non-native English writers. We describe four design principles that are relevant for correcting all of these errors, analyze the system along these dimensions, and show how each of these dimensions contributes to the performance.


2020 ◽  
Vol 8 ◽  
pp. 634-646
Author(s):  
Jared Lichtarge ◽  
Chris Alberti ◽  
Shankar Kumar

Recent progress in the task of Grammatical Error Correction (GEC) has been driven by addressing data sparsity, both through new methods for generating large and noisy pretraining data and through the publication of small and higher-quality finetuning data in the BEA-2019 shared task. Building upon recent work in Neural Machine Translation (NMT), we make use of both kinds of data by deriving example-level scores on our large pretraining data based on a smaller, higher-quality dataset. In this work, we perform an empirical study to discover how to best incorporate delta-log-perplexity, a type of example scoring, into a training schedule for GEC. In doing so, we perform experiments that shed light on the function and applicability of delta-log-perplexity. Models trained on scored data achieve state- of-the-art results on common GEC test sets.


2014 ◽  
Author(s):  
Hwee Tou Ng ◽  
Siew Mei Wu ◽  
Ted Briscoe ◽  
Christian Hadiwinoto ◽  
Raymond Hendy Susanto ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document