scholarly journals Development of English Handwritten Recognition Using Deep Neural Network

Author(s):  
Teddy Surya Gunawan ◽  
Ahmad Fakhrur Razi Mohd Noor ◽  
Mira Kartiwi

Due to the advanced in GPU and CPU, in recent years, Deep Neural Network (DNN) becomes popular to be utilized both as feature extraction and classifier. This paper aims to develop offline handwritten recognition system using DNN. First, two popular English digits and letters database, i.e. MNIST and EMNIST, were selected to provide dataset for training and testing phase of DNN. Altogether, there are 10 digits [0-9] and 52 letters [a-z, A-Z]. The proposed DNN used stacked two autoencoder layers and one softmax layer. Recognition accuracy for English digits and letters is 97.7% and 88.8%, respectively. Performance comparison with other structure of neural networks revealed that the weighted average recognition rate for patternnet, feedforwardnet, and proposed DNN were 80.3%, 68.3%, and 90.4%, respectively. It shows that our proposed system is able to recognize handwritten English digits and letters with high accuracy.

2002 ◽  
Vol 14 (01) ◽  
pp. 12-19 ◽  
Author(s):  
DUU-TONG FUH ◽  
CHING-HSING LUO

The standard Morse code defines the tone ratio (dash/dot) and the silent ratio (dash-space/dotspace) as 3:1. Since human typing ratio can't keep this ratio precisely and the two ratios —tone ratio and silent ratio—are not equal, the Morse code can't be recognized automatically. The requirement of the standard ratio is difficult to satisfy even for an ordinary person. As for the unstable Morse code typing pattern, the auto-recognition algorithms in the literature are not good enough in applications. The disabled persons usually have difficulty in maintaining a stable typing speeds and typing ratios, we therefore adopted an Expert-Gating neural network model to implement in single chip and recognize online unstable Morse codes. Also, we used another method—a linear back propagation recalling algorithm, to implement in single chip and recognize unstable Morse codes. From three person tests: Test one is a cerebral palsy; Test two is a beginner: Test three is a skilled expert, we have the results: in the experiment of test one, we have 91.15% (use 6 characters average time series as thresholds) and 91.54% (learning 26 characters) online average recognition rate; test two have 95.77% and 96.15%, and test three have 98.46% and 99.23% respectively. As for linear back propagation recalling method online recognized rate, we have the results from test one: 92.31% online average recognition rate; test two: 96.15%; and test three 99.23% respectively. So, we concluded: The Expert-Gating neural network and the linear back propagation recalling algorithm have successfully overcome the difficulty of analyzing a severely online unstable Morse code time series and successfully implement in single chip to recognize online unstable Morse code.


2021 ◽  
Vol 9 (2) ◽  
pp. 73-84
Author(s):  
Md. Shahadat Hossain ◽  
Md. Anwar Hossain ◽  
AFM Zainul Abadin ◽  
Md. Manik Ahmed

The recognition of handwritten Bangla digit is providing significant progress on optical character recognition (OCR). It is a very critical task due to the similar pattern and alignment of handwriting digits. With the progress of modern research on optical character recognition, it is reducing the complexity of the classification task by several methods, a few problems encounter during recognition and wait to be solved with simpler methods. The modern emerging field of artificial intelligence is the Deep Neural Network, which promises a solid solution to these few handwritten recognition problems. This paper proposed a fine regulated deep neural network (FRDNN) for the handwritten numeric character recognition problem that uses convolutional neural network (CNN) models with regularization parameters which makes the model generalized by preventing the overfitting. This paper applied Traditional Deep Neural Network (TDNN) and Fine regulated deep neural network (FRDNN) models with a similar layer experienced on BanglaLekha-Isolated databases and the classification accuracies for the two models were 96.25% and 96.99%, respectively over 100 epochs. The network performance of the FRDNN model on the BanglaLekha-Isolated digit dataset was more robust and accurate than the TDNN model and depend on experimentation. Our proposed method is obtained a good recognition accuracy compared with other existing available methods.


2021 ◽  
Vol 2021 ◽  
pp. 1-6
Author(s):  
Xueyan Chen ◽  
Xiaofei Zhong

In order to help pathologists quickly locate the lesion area, improve the diagnostic efficiency, and reduce missed diagnosis, a convolutional neural network algorithm for the optimization of emergency nursing rescue efficiency of critical patients was proposed. Specifically, three convolution layers and convolution kernels of different sizes are used to extract the features of patients’ posture behavior, and the classifier of patients’ posture behavior recognition system is used to learn the feature information by capturing the nonlinear relationship between the features to achieve accurate classification. By testing the accuracy of patient posture behavior feature extraction, the recognition rate of a certain action, and the average recognition rate of all actions in the patient body behavior recognition system, it is proved that the convolution neural network algorithm can greatly improve the efficiency of emergency nursing. The algorithm is applied to the patient posture behavior detection system, so as to realize the identification and monitoring of patients and improve the level of intelligent medical care. Finally, the open source framework platform is used to test the patient behavior detection system. The experimental results show that the larger the test data set is, the higher the accuracy of patient posture behavior feature extraction is, and the average recognition rate of patient posture behavior category is 97.6%, thus verifying the effectiveness and correctness of the system, to prove that the convolutional neural network algorithm has a very large improvement of emergency nursing rescue efficiency.


2021 ◽  
Vol 11 (11) ◽  
pp. 4758
Author(s):  
Ana Malta ◽  
Mateus Mendes ◽  
Torres Farinha

Maintenance professionals and other technical staff regularly need to learn to identify new parts in car engines and other equipment. The present work proposes a model of a task assistant based on a deep learning neural network. A YOLOv5 network is used for recognizing some of the constituent parts of an automobile. A dataset of car engine images was created and eight car parts were marked in the images. Then, the neural network was trained to detect each part. The results show that YOLOv5s is able to successfully detect the parts in real time video streams, with high accuracy, thus being useful as an aid to train professionals learning to deal with new equipment using augmented reality. The architecture of an object recognition system using augmented reality glasses is also designed.


2020 ◽  
Vol 11 (1) ◽  
pp. 10
Author(s):  
Muchun Su ◽  
Diana Wahyu Hayati ◽  
Shaowu Tseng ◽  
Jiehhaur Chen ◽  
Hsihsien Wei

Health care for independently living elders is more important than ever. Automatic recognition of their Activities of Daily Living (ADL) is the first step to solving the health care issues faced by seniors in an efficient way. The paper describes a Deep Neural Network (DNN)-based recognition system aimed at facilitating smart care, which combines ADL recognition, image/video processing, movement calculation, and DNN. An algorithm is developed for processing skeletal data, filtering noise, and pattern recognition for identification of the 10 most common ADL including standing, bending, squatting, sitting, eating, hand holding, hand raising, sitting plus drinking, standing plus drinking, and falling. The evaluation results show that this DNN-based system is suitable method for dealing with ADL recognition with an accuracy rate of over 95%. The findings support the feasibility of this system that is efficient enough for both practical and academic applications.


2020 ◽  
Vol 14 ◽  
Author(s):  
Stephanie Haro ◽  
Christopher J. Smalt ◽  
Gregory A. Ciccarelli ◽  
Thomas F. Quatieri

Many individuals struggle to understand speech in listening scenarios that include reverberation and background noise. An individual's ability to understand speech arises from a combination of peripheral auditory function, central auditory function, and general cognitive abilities. The interaction of these factors complicates the prescription of treatment or therapy to improve hearing function. Damage to the auditory periphery can be studied in animals; however, this method alone is not enough to understand the impact of hearing loss on speech perception. Computational auditory models bridge the gap between animal studies and human speech perception. Perturbations to the modeled auditory systems can permit mechanism-based investigations into observed human behavior. In this study, we propose a computational model that accounts for the complex interactions between different hearing damage mechanisms and simulates human speech-in-noise perception. The model performs a digit classification task as a human would, with only acoustic sound pressure as input. Thus, we can use the model's performance as a proxy for human performance. This two-stage model consists of a biophysical cochlear-nerve spike generator followed by a deep neural network (DNN) classifier. We hypothesize that sudden damage to the periphery affects speech perception and that central nervous system adaptation over time may compensate for peripheral hearing damage. Our model achieved human-like performance across signal-to-noise ratios (SNRs) under normal-hearing (NH) cochlear settings, achieving 50% digit recognition accuracy at −20.7 dB SNR. Results were comparable to eight NH participants on the same task who achieved 50% behavioral performance at −22 dB SNR. We also simulated medial olivocochlear reflex (MOCR) and auditory nerve fiber (ANF) loss, which worsened digit-recognition accuracy at lower SNRs compared to higher SNRs. Our simulated performance following ANF loss is consistent with the hypothesis that cochlear synaptopathy impacts communication in background noise more so than in quiet. Following the insult of various cochlear degradations, we implemented extreme and conservative adaptation through the DNN. At the lowest SNRs (<0 dB), both adapted models were unable to fully recover NH performance, even with hundreds of thousands of training samples. This implies a limit on performance recovery following peripheral damage in our human-inspired DNN architecture.


2012 ◽  
Vol 2012 ◽  
pp. 1-16 ◽  
Author(s):  
Khader Mohammad ◽  
Sos Agaian

Text embedded in an image contains useful information for applications in the medical, industrial, commercial, and research fields. While many systems have been designed to correctly identify text in images, no work addressing the recognition of degraded text on clear plastic has been found. This paper posits novel methods and an apparatus for extracting text from an image with the practical assumption: (a) poor background contrast, (b) white, curved, and/or differing fonts or character width between sets of images, (c) dotted text printed on curved reflective material, and/or (d) touching characters. Methods were evaluated using a total of 100 unique test images containing a variety of texts captured from water bottles. These tests averaged a processing time of ~10 seconds (using MATLAB R2008A on an HP 8510 W with 4 G of RAM and 2.3 GHz of processor speed), and experimental results yielded an average recognition rate of 90 to 93% using customized systems generated by the proposed development.


2021 ◽  
Vol 102 ◽  
pp. 04009
Author(s):  
Naoto Ageishi ◽  
Fukuchi Tomohide ◽  
Abderazek Ben Abdallah

Hand gestures are a kind of nonverbal communication in which visible bodily actions are used to communicate important messages. Recently, hand gesture recognition has received significant attention from the research community for various applications, including advanced driver assistance systems, prosthetic, and robotic control. Therefore, accurate and fast classification of hand gesture is required. In this research, we created a deep neural network as the first step to develop a real-time camera-only hand gesture recognition system without electroencephalogram (EEG) signals. We present the system software architecture in a fair amount of details. The proposed system was able to recognize hand signs with an accuracy of 97.31%.


2021 ◽  
Author(s):  
Wael Alnahari

Abstract In this paper, I proposed an iris recognition system by using deep learning via neural networks (CNN). Although CNN is used for machine learning, the recognition is achieved by building a non-trained CNN network with multiple layers. The main objective of the code the test pictures’ category (aka person name) with a high accuracy rate after having extracted enough features from training pictures of the same category which are obtained from a that I added to the code. I used IITD iris which included 10 iris pictures for 223 people.


Sign in / Sign up

Export Citation Format

Share Document