Text Recognition in Moving Vehicles using Deep learning Neural Networks

Author(s):  
K.T. Ilayarajaa ◽  
V. Vijayakumar ◽  
M. Sugadev ◽  
T. Ravi
2020 ◽  
Vol 69 (1) ◽  
pp. 378-383
Author(s):  
T.A. Nurmukhanov ◽  
◽  
B.S. Daribayev ◽  

Using neural networks, various variations of the classification of objects can be performed. Neural networks are used in many areas of recognition. A big area in this area is text recognition. The paper considers the optimal way to build a network for text recognition, the use of optimal methods for activation functions, and optimizers. Also, the article checked the correctness of text recognition with different optimization methods. This article is devoted to the analysis of convolutional neural networks. In the article, a convolutional neural network model will be trained with a teacher. Teaching with a teacher is a type of training for neural networks in which you provide the input data and the desired result, that is, the student looking at the input data will understand that you need to strive for the result that was provided to him.


2022 ◽  
pp. 161-219
Author(s):  
Chi-Hsuan Huang ◽  
Yu Sun ◽  
Chiou-Shana Fuh

In this chapter, an AI (artificial intelligence) solution for LPR (license plate recognition) on moving vehicles is proposed. The license plates in images captured with cameras on moving vehicles have unpredictable distortion and various illumination which make traditional machine vision algorithms unable to recognize the numbers correctly. Therefore, deep learning is leveraged to recognize license plate in such challenging conditions for better recognition accuracy. Additionally, lightweight neural networks are chosen since the power supply of scooter is quite limited. A two-stage method is presented to recognize license plate. First, the license plates in captured images are detected using CNN (convolutional neural network) model and the rotation of the detected license plates are corrected. Subsequently, the characters are recognized as upper-case format (A-Z) and digits (0-9) with second CNN model. Experimental results show that the system achieves 95.7% precision and 95% recall at high speed during the daytime.


Handwritten text recognition is a laborious task because humans can write a similar message in numerous ways or due to huge diversity in individual’s style of writing. The performance of text recognition systems implemented as neural networks has better results and accuracy than normal traditional classifiers. In this paper we explore the methods used to recognize and detect handwritten text or words in different languages. The major method used to recognize text is the Convolutional neural network (CNN) as a deep learning classifier. The other techniques used are Recurrent Neural Network (RNN) and a custom developed model called deep-writer, which is a variant of CNN architecture.


2020 ◽  
Author(s):  
Dean Sumner ◽  
Jiazhen He ◽  
Amol Thakkar ◽  
Ola Engkvist ◽  
Esben Jannik Bjerrum

<p>SMILES randomization, a form of data augmentation, has previously been shown to increase the performance of deep learning models compared to non-augmented baselines. Here, we propose a novel data augmentation method we call “Levenshtein augmentation” which considers local SMILES sub-sequence similarity between reactants and their respective products when creating training pairs. The performance of Levenshtein augmentation was tested using two state of the art models - transformer and sequence-to-sequence based recurrent neural networks with attention. Levenshtein augmentation demonstrated an increase performance over non-augmented, and conventionally SMILES randomization augmented data when used for training of baseline models. Furthermore, Levenshtein augmentation seemingly results in what we define as <i>attentional gain </i>– an enhancement in the pattern recognition capabilities of the underlying network to molecular motifs.</p>


2019 ◽  
Vol 277 ◽  
pp. 02024 ◽  
Author(s):  
Lincan Li ◽  
Tong Jia ◽  
Tianqi Meng ◽  
Yizhe Liu

In this paper, an accurate two-stage deep learning method is proposed to detect vulnerable plaques in ultrasonic images of cardiovascular. Firstly, a Fully Convonutional Neural Network (FCN) named U-Net is used to segment the original Intravascular Optical Coherence Tomography (IVOCT) cardiovascular images. We experiment on different threshold values to find the best threshold for removing noise and background in the original images. Secondly, a modified Faster RCNN is adopted to do precise detection. The modified Faster R-CNN utilize six-scale anchors (122,162,322,642,1282,2562) instead of the conventional one scale or three scale approaches. First, we present three problems in cardiovascular vulnerable plaque diagnosis, then we demonstrate how our method solve these problems. The proposed method in this paper apply deep convolutional neural networks to the whole diagnostic procedure. Test results show the Recall rate, Precision rate, IoU (Intersection-over-Union) rate and Total score are 0.94, 0.885, 0.913 and 0.913 respectively, higher than the 1st team of CCCV2017 Cardiovascular OCT Vulnerable Plaque Detection Challenge. AP of the designed Faster RCNN is 83.4%, higher than conventional approaches which use one-scale or three-scale anchors. These results demonstrate the superior performance of our proposed method and the power of deep learning approaches in diagnose cardiovascular vulnerable plaques.


Sensors ◽  
2020 ◽  
Vol 20 (6) ◽  
pp. 1579
Author(s):  
Dongqi Wang ◽  
Qinghua Meng ◽  
Dongming Chen ◽  
Hupo Zhang ◽  
Lisheng Xu

Automatic detection of arrhythmia is of great significance for early prevention and diagnosis of cardiovascular disease. Traditional feature engineering methods based on expert knowledge lack multidimensional and multi-view information abstraction and data representation ability, so the traditional research on pattern recognition of arrhythmia detection cannot achieve satisfactory results. Recently, with the increase of deep learning technology, automatic feature extraction of ECG data based on deep neural networks has been widely discussed. In order to utilize the complementary strength between different schemes, in this paper, we propose an arrhythmia detection method based on the multi-resolution representation (MRR) of ECG signals. This method utilizes four different up to date deep neural networks as four channel models for ECG vector representations learning. The deep learning based representations, together with hand-crafted features of ECG, forms the MRR, which is the input of the downstream classification strategy. The experimental results of big ECG dataset multi-label classification confirm that the F1 score of the proposed method is 0.9238, which is 1.31%, 0.62%, 1.18% and 0.6% higher than that of each channel model. From the perspective of architecture, this proposed method is highly scalable and can be employed as an example for arrhythmia recognition.


Sign in / Sign up

Export Citation Format

Share Document