Offline Handwritten Gurumukhi Character Recognition System Using Convolution Neural Network

2019 ◽  
Vol 16 (10) ◽  
pp. 4164-4169
Author(s):  
Sheifali Gupta ◽  
Udit Jindal ◽  
Deepali Gupta ◽  
Rupesh Gupta

A lot of literature is available on the recognition of handwriting on scripts other than Indians, but the number of articles related to Indian scripts recognition such as Gurumukhi are much less. Gurumukhi is a religion-specific language that ranks 14th frequently spoken language in all languages of the world. In Gurumukhi script, some characters are alike to each other which makes recognition task very difficult. Therefore this article presents a novel approach for Gurumukhi character. This article lays emphasis on convolutional neural networks (CNN), which intend to obtain the features of given data samples and then its mapping is being performed to the right observation. In this approach, a dataset has been prepared for 10 Gurumukhi characters. The proposed methodology obtains a recognition accuracy of 99.34% on Gurumukhi characters images without making use of any post-processing method.

2022 ◽  
Vol 12 (2) ◽  
pp. 853
Author(s):  
Cheng-Jian Lin ◽  
Yu-Cheng Liu ◽  
Chin-Ling Lee

In this study, an automatic receipt recognition system (ARRS) is developed. First, a receipt is scanned for conversion into a high-resolution image. Receipt characters are automatically placed into two categories according to the receipt characteristics: printed and handwritten characters. Images of receipts with these characters are preprocessed separately. For handwritten characters, template matching and the fixed features of the receipts are used for text positioning, and projection is applied for character segmentation. Finally, a convolutional neural network is used for character recognition. For printed characters, a modified You Only Look Once (version 4) model (YOLOv4-s) executes precise text positioning and character recognition. The proposed YOLOv4-s model reduces downsampling, thereby enhancing small-object recognition. Finally, the system produces recognition results in a tax declaration format, which can upload to a tax declaration system. Experimental results revealed that the recognition accuracy of the proposed system was 80.93% for handwritten characters. Moreover, the YOLOv4-s model had a 99.39% accuracy rate for printed characters; only 33 characters were misjudged. The recognition accuracy of the YOLOv4-s model was higher than that of the traditional YOLOv4 model by 20.57%. Therefore, the proposed ARRS can considerably improve the efficiency of tax declaration, reduce labor costs, and simplify operating procedures.


2018 ◽  
Vol 7 (3.4) ◽  
pp. 90 ◽  
Author(s):  
Mandeep Singh ◽  
Karun Verma ◽  
Bob Gill ◽  
Ramandeep Kaur

Online handwriting character recognition is gaining attention from the researchers across the world because with the advent of touch based devices, a more natural way of communication is being explored. Stroke based online recognition system is proposed in this paper for a very complex Gurmukhi script. In this effort, recognition for 35 basic characters of Gurmukhi script has been implemented on the dataset of 2019 Gurmukhi samples. For this purpose, 32 stroke classes have been considered. Three types of features have been extracted. Hybrid of these features has been proposed in this paper to train the classification models. For stroke classification, three different classifiers namely, KNN, MLP and SVM are used and compared to evaluate the effectiveness of these models. A very promising “stroke recognition rate” of 94% by KNN, 95.04% by MLP and 95.04% by SVM has been obtained.  


2013 ◽  
Vol 333-335 ◽  
pp. 883-887
Author(s):  
Yong Xia ◽  
Zhi Bo Yang ◽  
Kuan Quan Wang

t is quite constrained for us to use some other input devices to communicate with computers. In this paper, we integrate human-computer interaction technologies with handwritten Chinese character recognition strategies using depth image information provided by Kinect sensor to realize an unconstrained handwritten character recognition system, which only uses our hand as input device. We predefine several hand gestures as instructions, and for the recognition of these hand gestures, we calculate the contour and fingertips of the hand used for writing using depth image taken by Kinect. By mimicking the functionalities of the computer mouse only using our hands, we can write freely in the air and get the original character image. After Gaussian blurring and normalization, we adopt some classic handwritten character recognition schemes to accomplish the recognition task. Experiments show that the system gives a good result.


1999 ◽  
Vol 09 (06) ◽  
pp. 545-561 ◽  
Author(s):  
HSIN-CHIA FU ◽  
Y. Y. XU ◽  
H. Y. CHANG

Recognition of similar (confusion) characters is a difficult problem in optical character recognition (OCR). In this paper, we introduce a neural network solution that is capable of modeling minor differences among similar characters, and is robust to various personal handwriting styles. The Self-growing Probabilistic Decision-based Neural Network (SPDNN) is a probabilistic type neural network, which adopts a hierarchical network structure with nonlinear basis functions and a competitive credit-assignment scheme. Based on the SPDNN model, we have constructed a three-stage recognition system. First, a coarse classifier determines a character to be input to one of the pre-defined subclasses partitioned from a large character set, such as Chinese mixed with alphanumerics. Then a character recognizer determines the input image which best matches the reference character in the subclass. Lastly, the third module is a similar character recognizer, which can further enhance the recognition accuracy among similar or confusing characters. The prototype system has demonstrated a successful application of SPDNN to similar handwritten Chinese recognition for the public database CCL/HCCR1 (5401 characters × 200 samples). Regarding performance, experiments on the CCL/HCCR1 database produced 90.12% recognition accuracy with no rejection, and 94.11% accuracy with 6.7% rejection, respectively. This recognition accuracy represents about 4% improvement on the previously announced performance.5,11 As to processing speed, processing before recognition (including image preprocessing, segmentation, and feature extraction) requires about one second for an A4 size character image, and recognition consumes approximately 0.27 second per character on a Pentium-100 based personal computer, without use of any hardware accelerator or co-processor.


2020 ◽  
Author(s):  
Nishatul Majid

This dissertation presents a flexible and robust offline handwriting recognition system which is tested on the Bangla and Korean scripts. Offline handwriting recognition is one of the most challenging and yet to be solved problems in machine learning. While a few popular scripts (like Latin) have received a lot of attention, many other widely used scripts (like Bangla) have seen very little progress. Features such as connectedness and vowels structured as diacritics make it a challenging script to recognize. A simple and robust design for offline recognition is presented which not only works reliably, but also can be used for almost any alphabetic writing system. The framework has been rigorously tested for Bangla and demonstrated how it can be transformed to apply to other scripts through experiments on the Korean script whose two-dimensional arrangement of characters makes it a challenge to recognize. The base of this design is a character spotting network which detects the location of different script elements (such as characters, diacritics) from an unsegmented word image. A transcript is formed from the detected classes based on their corresponding location information. This is the first reported lexicon-free offline recognition system for Bangla and achieves a Character Recognition Accuracy (CRA) of 94.8%. This is also one of the most flexible architectures ever presented. Recognition of Korean was achieved with a 91.2% CRA. Also, a powerful technique of autonomous tagging was developed which can drastically reduce the effort of preparing a dataset for any script. The combination of the character spotting method and the autonomous tagging brings the entire offline recognition problem very close to a singular solution. Additionally, a database named the Boise State Bangla Handwriting Dataset was developed. This is one of the richest offline datasets currently available for Bangla and this has been made publicly accessible to accelerate the research progress. Many other tools were developed and experiments were conducted to more rigorously validate this framework by evaluating the method against external datasets (CMATERdb 1.1.1, Indic Word Dataset and REID2019: Early Indian Printed Documents). Offline handwriting recognition is an extremely promising technology and the outcome of this research moves the field significantly ahead.


Author(s):  
Binod Kumar Prasad

Purpose: Lines and Curves are important parts of characters in any script. Features based on lines and curves go a long way to characterize an individual character as well as differentiate similar-looking characters. The present paper proposes an English numerals recognition system using feature elements obtained from the novel and efficient coding of the curves and local slopes. The purpose of this paper is to recognize English numerals efficiently to develop a reliable Optical Character recognition system. Methodology: K-Nearest Neighbour classification technique has been implemented on a global database MNIST to get an overall recognition accuracy rate of 96.7 %, which is competitive to other reported works in literature. Distance features and slope features are extracted from pre-processed images. The feature elements from training images are used to train K-Nearest-Neighbour classifier and those from test images have been used to classify them. Main Findings: The findings of the current paper can be used in Optical Character Recognition (OCR) of alphanumeric characters of any language, automatic reading of amount on bank cheque, address written on envelops, etc. Implications: Due to the similarity in structures of some numerals like 2, 3, and 8, the system produces respectively lower recognition accuracy rates for them. Novelty: The ways of finding distance and slope features to differentiate the curves in the structure of English Numerals is the novelty of this work.


Author(s):  
SURESH KUMAR D S ◽  
AJAY KUMAR B R ◽  
K SRINIVASA KALYAN

Handwriting recognition has been one of the active and challenging research areas in the field of pattern recognition. It has numerous applications which include, reading aid for blind, bank cheques and conversion of any hand written document into structural text form[1]. As there are no sufficient number of works on Indian language character recognition especially Kannada script among 15 major scripts in India[2].In this paper an attempt is made to recognize handwritten Kannada characters using Feed Forward neural networks. A handwritten kannada character is resized into 20x30 pixel.The resized character is used for training the neural network. Once the training process is completed the same character is given as input to the neural network with different set of neurons in hidden layer and their recognition accuracy rate for different kannada characters has been calculated and compared. The results show that the proposed system yields good recognition accuracy rates comparable to that of other handwritten character recognition systems.


offline handwritten character recognition system has been a challenge for Indian scripts, especially for South Indian languages. Huge number of characters of local languages including alphabets, consonants and composite characters make the recognition system more complicated. A good recognition system for subset of Tamil script, a famous South Indian script, is proposed in this work. Variable length feature vector is extracted from the thinned character image. This extracted feature is given to a novel simple classification algorithm which works based on probability. A subset of Tamil script, 20 character classes, is considered for experiment. The samples were taken from HP Labs dataset for Tamil language and a recognition accuracy of 88.15% has been produced.


Sign in / Sign up

Export Citation Format

Share Document