Intriguing Aspects of Oriental Languages

Author(s):  
Ching Y. Suen ◽  
Shunji Mori ◽  
Hae-Chang Rim ◽  
Patrick S. P. Wang

This paper includes a description of 3 affiliated oriental languages: Chinese, Japanese, and Korean. It includes a description of the origins of these 3 languages and the inter-relationship among them. Drawn from the viewpoints of several experienced researchers in the field of OCR (Optical Character Recognition) and computational linguistics, it attempts to bring out the intriguing aspects of these 3 ideographic languages, including the formation and composition of pictograms, special features, learning, understanding, contextual information, and recognition of characters and words, and their relations to poetic expressions and pattern recognition techniques. Numerous references are given and comments on future trends are also presented.

Author(s):  
Shubhankar Sharma ◽  
Vatsala Arora

The study of character research is an active area for research as it pertains a lot of challenges. Various pattern recognition techniques are being used every day. As there are so many writing styles available, development of OCR (Optical Character Recognition) for handwritten text is difficult. Therefore, several measures have to be taken to improve the recognition process so that the burden of computation can be decreased and the accuracy for pattern recognition can be increased. The main objective of this review was to recognize and analyze handwritten document images. In this paper, we present a scheme to identify different Indian scripts like Devanagari and Gurumukhi.


1994 ◽  
Vol 04 (01) ◽  
pp. 193-207 ◽  
Author(s):  
VADIM BIKTASHEV ◽  
VALENTIN KRINSKY ◽  
HERMANN HAKEN

The possibility of using nonlinear media as a highly parallel computation tool is discussed, specifically for image classification and recognition. Some approaches of this type are known, that are based on stationary dissipative structures which can “measure” scalar products of images. In this paper, we exploit the analogy between binary images and point sets, and use the Hausdorff metrics for comparing the images. It does not require the measure at all, and is based only on the metrics of the space whose subsets we consider. In addition to Hausdorff distance, we suggest a new “nonlinear” version of this distance for comparison of images, called “autowave” distance. This distance can be calculated very easily and yields some additional advantages for pattern recognition (e.g. noise tolerance). The method was illustrated for the problem of machine reading (Optical Character Recognition). It was compared with some famous OCR programs for PC. On a medium quality xerocopy of a journal page, in the same conditions of learning and recognition, the autowave approach resulted in much fewer mistakes. The method can be realized using only one chip with simple uniform connection of the elements. In this case, it yields an increase in computation speed of several orders of magnitude.


Author(s):  
Husni Al-Muhtaseb ◽  
Rami Qahwaji

Arabic text recognition is receiving more attentions from both Arabic and non-Arabic-speaking researchers. This chapter provides a general overview of the state-of-the-art in Arabic Optical Character Recognition (OCR) and the associated text recognition technology. It also investigates the characteristics of the Arabic language with respect to OCR and discusses related research on the different phases of text recognition including: pre-processing and text segmentation, common feature extraction techniques, classification methods and post-processing techniques. Moreover, the chapter discusses the available databases for Arabic OCR research and lists the available commercial Software. Finally, it explores the challenges related to Arabic OCR and discusses possible future trends.


Research is deliberately going on in the field of pattern recognition. New ideas are developed and implemented in this field throughout the globe. Optical Character Recognition (OCR) is one of the inseparable applications of Pattern Recognition. Though extensive research is already reported in this field, but multilingual Optical Character Recognition is the most challenging aspect which is still, the need of the hour. Myriads of researchers are digging the information to gather the best solutions for the recognition purpose. In this research paper, we are purposing the steps for the recognition of Devanagari and English scripts simultaneously occurring in the documents. A new approach of segmentation and splitting the characters of both the scripts is also introduced for the benefits of researchers. Most commonly in the documents containing English and Devanagari scripts, English characters are already separated, the challenge is to separate the Devanagari characters. Algorithm to implement the challenging aspect to segment the Devanagari and Roman scripts simultaneously is also implemented in the present paper.


License plate recognition system plays very important role in various security aspects which includes entry monitoring of a particular vehicle in commercial complex, traffic monitoring , identification of threats and many more. In past few years many different methods has been adopted for license plate recognition system but still there is little more chance to work on real time difficulties which come across while license plate recognition like speed of vehicle, angle of license plate in picture, background of picture or color contrast of image, reflection on the license plate and so on. The combination of object detection, image processing, and pattern recognition are used to fulfill this application. In the proposed architecture , system will capture a small video and using Google's OCR(Optical Character Recognition) system will recognize license number, if that number get found in database gate will get open with the help of Arduino Uno.


Robotics ◽  
2021 ◽  
Vol 10 (4) ◽  
pp. 110
Author(s):  
Daniel Dworakowski ◽  
Christopher Thompson ◽  
Michael Pham-Hung ◽  
Goldie Nejat

Grocery shoppers must negotiate cluttered, crowded, and complex store layouts containing a vast variety of products to make their intended purchases. This complexity may prevent even experienced shoppers from finding their grocery items, consuming a lot of their time and resulting in monetary loss for the store. To address these issues, we present a generic grocery robot architecture for the autonomous search and localization of products in crowded dynamic unknown grocery store environments using a unique context Simultaneous Localization and Mapping (contextSLAM) method. The contextSLAM method uniquely creates contextually rich maps through the online fusion of optical character recognition and occupancy grid information to locate products and aid in robot localization in an environment. The novelty of our robot architecture is in its ability to intelligently use geometric and contextual information within the context map to direct robot exploration in order to localize products in unknown environments in the presence of dynamic people. Extensive experiments were conducted with a mobile robot to validate the overall architecture and contextSLAM, including in a real grocery store. The results of the experiments showed that our architecture was capable of searching for and localizing all products in various grocery lists in different unknown environments.


2012 ◽  
Vol 424-425 ◽  
pp. 1107-1111
Author(s):  
Fu Cheng You ◽  
Ying Jie Liu

For the purpose of information management on postmark according to the date, the paper put forward a method of postmark date recognition based on machine vision, which could meet the demands of personal postmark collectors. On the basis of the relative theories of machine vision, image processing and pattern recognition, the overall process is introduced in the paper from postmark image acquisition to date recognition. Firstly, threshold method is used to generate binary image from smoothed postmark image. So region of date numbers could be extracted from binary image according to different region features. Then regions of date numbers which are connected or broken could be processed through mathematical morphology of binary image. Individual regions of date numbers are obtained for recognition. Finally, classification and pattern recognition based on support vector machine make date numbers classified and date recognition is implemented correctly


1997 ◽  
Vol 9 (1-3) ◽  
pp. 58-77
Author(s):  
Vitaly Kliatskine ◽  
Eugene Shchepin ◽  
Gunnar Thorvaldsen ◽  
Konstantin Zingerman ◽  
Valery Lazarev

In principle, printed source material should be made machine-readable with systems for Optical Character Recognition, rather than being typed once more. Offthe-shelf commercial OCR programs tend, however, to be inadequate for lists with a complex layout. The tax assessment lists that assess most nineteenth century farms in Norway, constitute one example among a series of valuable sources which can only be interpreted successfully with specially designed OCR software. This paper considers the problems involved in the recognition of material with a complex table structure, outlining a new algorithmic model based on ‘linked hierarchies’. Within the scope of this model, a variety of tables and layouts can be described and recognized. The ‘linked hierarchies’ model has been implemented in the ‘CRIPT’ OCR software system, which successfully reads tables with a complex structure from several different historical sources.


2020 ◽  
Vol 2020 (1) ◽  
pp. 78-81
Author(s):  
Simone Zini ◽  
Simone Bianco ◽  
Raimondo Schettini

Rain removal from pictures taken under bad weather conditions is a challenging task that aims to improve the overall quality and visibility of a scene. The enhanced images usually constitute the input for subsequent Computer Vision tasks such as detection and classification. In this paper, we present a Convolutional Neural Network, based on the Pix2Pix model, for rain streaks removal from images, with specific interest in evaluating the results of the processing operation with respect to the Optical Character Recognition (OCR) task. In particular, we present a way to generate a rainy version of the Street View Text Dataset (R-SVTD) for "text detection and recognition" evaluation in bad weather conditions. Experimental results on this dataset show that our model is able to outperform the state of the art in terms of two commonly used image quality metrics, and that it is capable to improve the performances of an OCR model to detect and recognise text in the wild.


Sign in / Sign up

Export Citation Format

Share Document