Automatic Error Detection and Correction of Text: The State of the Art

Author(s):  
Zhang Yang ◽  
Zhao Xiaobing
2018 ◽  
Vol 111 (1) ◽  
pp. 97-112
Author(s):  
Tim vor der Brück

Abstract Rule-based natural language generation denotes the process of converting a semantic input structure into a surface representation by means of a grammar. In the following, we assume that this grammar is handcrafted and not automatically created for instance by a deep neural network. Such a grammar might comprise of a large set of rules. A single error in these rules can already have a large impact on the quality of the generated sentences, potentially causing even a complete failure of the entire generation process. Searching for errors in these rules can be quite tedious and time-consuming due to potentially complex and recursive dependencies. This work proposes a statistical approach to recognizing errors and providing suggestions for correcting certain kinds of errors by cross-checking the grammar with the semantic input structure. The basic assumption is the correctness of the latter, which is usually a valid hypothesis due to the fact that these input structures are often automatically created. Our evaluation reveals that in many cases an automatic error detection and correction is indeed possible.


Electronics ◽  
2018 ◽  
Vol 7 (10) ◽  
pp. 258 ◽  
Author(s):  
Abdus Hassan ◽  
Umar Afzaal ◽  
Tooba Arifeen ◽  
Jeong Lee

Recently, concurrent error detection enabled through invariant relationships between different wires in a circuit has been proposed. Because there are many such implications in a circuit, selection strategies have been developed to select the most valuable implications for inclusion in the checker hardware such that a sufficiently high probability of error detection ( P d e t e c t i o n ) is achieved. These algorithms, however, due to their heuristic nature cannot guarantee a lossless P d e t e c t i o n . In this paper, we develop a new input-aware implication selection algorithm with the help of ATPG which minimizes loss on P d e t e c t i o n . In our algorithm, the detectability of errors for each candidate implication is carefully evaluated using error prone vectors. The evaluation results are then utilized to select the most efficient candidates for achieving optimal P d e t e c t i o n . The experimental results on 15 representative combinatorial benchmark circuits from the MCNC benchmarks suite show that the implications selected from our algorithm achieve better P d e t e c t i o n in comparison to the state of the art. The proposed method also offers better performance, up to 41.10%, in terms of the proposed impact-level metric, which is the ratio of achieved P d e t e c t i o n to the implication count.


Author(s):  
CHUEN-MIN HUANG ◽  
MEI-CHEN WU ◽  
CHING-CHE CHANG

Misspelling and misconception resulting from similar pronunciation appears frequently in Chinese texts. Without double check-up, this situation will be getting worse even with the help of Chinese input editor. It is hoped that the quality of Chinese writing would be enhanced if an effective automatic error detection and correction mechanism is embedded in text editor. Therefore, the burden of manpower to proofread shall be released. Until recently, researches in automatic error detection and correction of Chinese text have undergone many challenges and suffered from bad performance compared with that of Western text. In view of the prominent phenomenon in Chinese writing problem, this study proposes a learning model based on Chinese phonemic alphabets. The experimental results demonstrate that this model is effective in finding out misspellings and further improves detection and correction rate.


2019 ◽  
Vol 26 (3) ◽  
pp. 211-218 ◽  
Author(s):  
Chris J Lu ◽  
Alan R Aronson ◽  
Sonya E Shooshan ◽  
Dina Demner-Fushman

Abstract Objective Automated understanding of consumer health inquiries might be hindered by misspellings. To detect and correct various types of spelling errors in consumer health questions, we developed a distributable spell-checking tool, CSpell, that handles nonword errors, real-word errors, word boundary infractions, punctuation errors, and combinations of the above. Methods We developed a novel approach of using dual embedding within Word2vec for context-dependent corrections. This technique was used in combination with dictionary-based corrections in a 2-stage ranking system. We also developed various splitters and handlers to correct word boundary infractions. All correction approaches are integrated to handle errors in consumer health questions. Results Our approach achieves an F1 score of 80.93% and 69.17% for spelling error detection and correction, respectively. Discussion The dual-embedding model shows a significant improvement (9.13%) in F1 score compared with the general practice of using cosine similarity with word vectors in Word2vec for context ranking. Our 2-stage ranking system shows a 4.94% improvement in F1 score compared with the best 1-stage ranking system. Conclusion CSpell improves over the state of the art and provides near real-time automatic misspelling detection and correction in consumer health questions. The software and the CSpell test set are available at https://umlslex.nlm.nih.gov/cSpell.


2009 ◽  
pp. 110-144
Author(s):  
Piotr Augustyniak ◽  
Ryszard Tadeusiewicz

This chapter defines the set of standard diagnostic parameters and metadata expected from cardiac examinations. Rest ECG, exercise ECG, and long-term recording techniques are compared with regard to method-appropriate hierarchies of diagnostic results. This summary is approaching the idea of high redundancy in the dataset influencing data transmission and database operations. As far as the paper record was concerned, these spare data were useful in the validation and correction of human errors. Nowadays, automatic error detection and correction codes are widely applied in systems for storage and transmission of digital data. Basic issues about DICOM and HL7, two widespread medical information interchange systems, are presented thereafter. These general-purpose systems integrate multi-modal medical data and offer specialized tools for the storage, retrieval, and management of data. Both standards originate from the efforts of standardizing the description of possibly wide aspects of patient-oriented digital data in the form of electronic health records. Certain aspects of data security are also considered here.


Sign in / Sign up

Export Citation Format

Share Document