Fast Recognition of Noisy Digits

1993 ◽  
Vol 5 (6) ◽  
pp. 885-892 ◽  
Author(s):  
Jeffrey N. Kidder ◽  
Daniel Seligson

We describe a hardware solution to a high-speed optical character recognition (OCR) problem. Noisy 15 × 10 binary images of machine written digits were processed and applied as input to Intel's Electrically Trainable Analog Neural Network (ETANN). In software simulation, we trained an 80 × 54 × 10 feedforward network using a modified version of backprop. We then downloaded the synaptic weights of the trained network to ETANN and tweaked them to account for differences between the simulation and the chip itself. The best recognition error rate was 0.9% in hardware with a 3.7% rejection rate on a 1000-character test set.

1979 ◽  
Vol 73 (10) ◽  
pp. 389-399
Author(s):  
Gregory L. Goodrich ◽  
Richard R. Bennett ◽  
William R. De L'aune ◽  
Harvey Lauer ◽  
Leonard Mowinski

This study was designed to assess the Kurzweil Reading Machine's ability to read three different type styles produced by five different means. The results indicate that the Kurzweil Reading Machines tested have different error rates depending upon the means of producing the copy and upon the type style used; there was a significant interaction between copy method and type style. The interaction indicates that some type styles are better read when the copy is made by one means rather than another. Error rates varied between less than one percent and more than twenty percent. In general, the user will find that high quality printed materials will be read with a relatively high level of accuracy, but as the quality of the material decreases, the number of errors made by the machine also increases. As this error rate increases, the user will find it increasingly difficult to understand the spoken output.


1994 ◽  
Vol 04 (01) ◽  
pp. 193-207 ◽  
Author(s):  
VADIM BIKTASHEV ◽  
VALENTIN KRINSKY ◽  
HERMANN HAKEN

The possibility of using nonlinear media as a highly parallel computation tool is discussed, specifically for image classification and recognition. Some approaches of this type are known, that are based on stationary dissipative structures which can “measure” scalar products of images. In this paper, we exploit the analogy between binary images and point sets, and use the Hausdorff metrics for comparing the images. It does not require the measure at all, and is based only on the metrics of the space whose subsets we consider. In addition to Hausdorff distance, we suggest a new “nonlinear” version of this distance for comparison of images, called “autowave” distance. This distance can be calculated very easily and yields some additional advantages for pattern recognition (e.g. noise tolerance). The method was illustrated for the problem of machine reading (Optical Character Recognition). It was compared with some famous OCR programs for PC. On a medium quality xerocopy of a journal page, in the same conditions of learning and recognition, the autowave approach resulted in much fewer mistakes. The method can be realized using only one chip with simple uniform connection of the elements. In this case, it yields an increase in computation speed of several orders of magnitude.


The process of an Optical Character Recognition (OCR) for ancient hand written documents or palm leaf manuscripts is done by means of four phases. The four phases are ‘line segmentation’, ‘word segmentation’, ‘character segmentation’, and ‘character recognition’. The colour image of palm leaf manuscripts are changed into binary images by using various pre-processing methods. The first phase of an OCR might break through the hurdles of touching lines and overlapping lines. The character recognition becomes futile when the line segmentation is erroneous. In Tamil language palm leaf manuscript recognition, there are only a handful of line segmentation methods. Moreover, the available methods are not viable to meet the required standards. This article is proposed to fill the lacuna in terms of the methods necessary for line segmentation in Tamil language document analysis. The method proposed compares its efficiency with the line segmentation algorithms work on binary images such as the Adaptive Partial Projection (APP) and A* Path Planning (A*PP). The tools and criteria of evaluation metrics are measured from ICDAR 2013 Handwriting Segmentation Contest.


2018 ◽  
pp. 707-734
Author(s):  
F. Daneshfar ◽  
W. Fathy ◽  
B. Alaqeband

Preprocessing is a very important part of cursive languages Optical Character Recognition (OCR) systems. Thus, baseline detection, which is one of the main parts of the preprocessing operation, plays a basic role on OCR systems; improvement on baseline detection could be absolutely useful for decreasing errors in recognition words. In this chapter, a metaheuristic- and mathematical-based algorithm is recommended, which has improved the baseline detection process in relation to the well-known baseline detection algorithms. The most important advantages of the proposed method are simplicity, high speed processing, and reliability. To test this novel solution, IFN/ENIT database, which is a well-known and attending database, is utilized. However, the proposed solution is reliable to any standard database of cursive language's OCR.


Author(s):  
F. Daneshfar ◽  
W. Fathy ◽  
B. Alaqeband

Preprocessing is a very important part of cursive languages Optical Character Recognition (OCR) systems. Thus, baseline detection, which is one of the main parts of the preprocessing operation, plays a basic role on OCR systems; improvement on baseline detection could be absolutely useful for decreasing errors in recognition words. In this chapter, a metaheuristic- and mathematical-based algorithm is recommended, which has improved the baseline detection process in relation to the well-known baseline detection algorithms. The most important advantages of the proposed method are simplicity, high speed processing, and reliability. To test this novel solution, IFN/ENIT database, which is a well-known and attending database, is utilized. However, the proposed solution is reliable to any standard database of cursive language's OCR.


2018 ◽  
Vol 246 ◽  
pp. 03040
Author(s):  
Jie Kong ◽  
Congying Wang

In recent years, although Optical Character Recognition (OCR) has made considerable progress, low-resolution text images commonly appearing in many scenarios may still cause errors in recognition. For this problem, the technique of Generative Adversarial Network in super-resolution processing is applied to enhance the resolution of low-quality text images in this study. The principle and the implementation in TensorFlow of this technique are introduced. On this basis, a system is proposed to perform the resolution enhancement and OCR for low-resolution text images. The experimental results indicate that this technique could significantly improve the accuracy, reduce the error rate and false rejection rate of low-resolution text images identification.


1997 ◽  
Vol 9 (1-3) ◽  
pp. 58-77
Author(s):  
Vitaly Kliatskine ◽  
Eugene Shchepin ◽  
Gunnar Thorvaldsen ◽  
Konstantin Zingerman ◽  
Valery Lazarev

In principle, printed source material should be made machine-readable with systems for Optical Character Recognition, rather than being typed once more. Offthe-shelf commercial OCR programs tend, however, to be inadequate for lists with a complex layout. The tax assessment lists that assess most nineteenth century farms in Norway, constitute one example among a series of valuable sources which can only be interpreted successfully with specially designed OCR software. This paper considers the problems involved in the recognition of material with a complex table structure, outlining a new algorithmic model based on ‘linked hierarchies’. Within the scope of this model, a variety of tables and layouts can be described and recognized. The ‘linked hierarchies’ model has been implemented in the ‘CRIPT’ OCR software system, which successfully reads tables with a complex structure from several different historical sources.


2020 ◽  
Vol 2020 (1) ◽  
pp. 78-81
Author(s):  
Simone Zini ◽  
Simone Bianco ◽  
Raimondo Schettini

Rain removal from pictures taken under bad weather conditions is a challenging task that aims to improve the overall quality and visibility of a scene. The enhanced images usually constitute the input for subsequent Computer Vision tasks such as detection and classification. In this paper, we present a Convolutional Neural Network, based on the Pix2Pix model, for rain streaks removal from images, with specific interest in evaluating the results of the processing operation with respect to the Optical Character Recognition (OCR) task. In particular, we present a way to generate a rainy version of the Street View Text Dataset (R-SVTD) for "text detection and recognition" evaluation in bad weather conditions. Experimental results on this dataset show that our model is able to outperform the state of the art in terms of two commonly used image quality metrics, and that it is capable to improve the performances of an OCR model to detect and recognise text in the wild.


Sign in / Sign up

Export Citation Format

Share Document