scholarly journals Recognition of Brahmi Words by Using Deep Convolutional Neural Network

Author(s):  
Neha Gautam ◽  
Soo See Chai ◽  
Jais Jose

Significant progress has made in pattern recognition technology. However, one obstacle that has not yet overcome is the recognition of words in the Brahmi script, specifically the identification of characters, compound characters, and word. This study proposes the use of the deep convolutional neural network with dropout to recognize the Brahmi words. This study also proposed a DCNN for Brahmi word recognition and a series of experiments are performed on standard Brahmi dataset. The practical operation of this method was systematically tested on accessible Brahmi image database, achieving 92.47% recognition rate by CNN with dropout respectively which is among the best while comparing with the ones reported in the literature for the same task.

2019 ◽  
Vol 9 (13) ◽  
pp. 2758 ◽  
Author(s):  
Mujtaba Husnain ◽  
Malik Muhammad Saad Missen ◽  
Shahzad Mumtaz ◽  
Muhammad Zeeshan Jhanidr ◽  
Mickaël Coustaty ◽  
...  

In the area of pattern recognition and pattern matching, the methods based on deep learning models have recently attracted several researchers by achieving magnificent performance. In this paper, we propose the use of the convolutional neural network to recognize the multifont offline Urdu handwritten characters in an unconstrained environment. We also propose a novel dataset of Urdu handwritten characters since there is no publicly-available dataset of this kind. A series of experiments are performed on our proposed dataset. The accuracy achieved for character recognition is among the best while comparing with the ones reported in the literature for the same task.


Author(s):  
G. Merlin Linda ◽  
G. Themozhi ◽  
Sudheer Reddy Bandi

In recent decades, gait recognition has garnered a lot of attention from the researchers in the IT era. Gait recognition signifies verifying or identifying the individuals by their walking style. Gait supports in surveillance system by identifying people when they are at a distance from the camera and can be used in numerous computer vision and surveillance applications. This paper proposes a stupendous Color-mapped Contour Gait Image (CCGI) for varying factors of Cross-View Gait Recognition (CVGR). The first contour in each gait image sequence is extracted using a Combination of Receptive Fields (CORF) contour tracing algorithm which extracts the contour image using Difference of Gaussians (DoG) and hysteresis thresholding. Moreover, hysteresis thresholding detects the weak edges from the total pixel information and provides more well-balanced smooth features compared to an absolute one. Second CCGI encodes the spatial and temporal information via color mapping to attain the regularized contour images with fewer outliers. Based on the front view of a human walking pattern, the appearance of cross-view variations would reduce drastically with respect to a change of view angles. This proposed work evaluates the performance analysis of CVGR using Deep Convolutional Neural Network (CNN) framework. CCGI is considered a gait feature for comparing and evaluating the robustness of our proposed model. Experiments conducted on CASIA-B database show the comparisons of previous methods with the proposed method and achieved 94.65% accuracy with a better recognition rate.


2020 ◽  
Vol 37 (12) ◽  
pp. 2197-2207
Author(s):  
Andrew Geiss ◽  
Joseph C. Hardin

AbstractSuper resolution involves synthetically increasing the resolution of gridded data beyond their native resolution. Typically, this is done using interpolation schemes, which estimate sub-grid-scale values from neighboring data, and perform the same operation everywhere regardless of the large-scale context, or by requiring a network of radars with overlapping fields of view. Recently, significant progress has been made in single-image super resolution using convolutional neural networks. Conceptually, a neural network may be able to learn relations between large-scale precipitation features and the associated sub-pixel-scale variability and outperform interpolation schemes. Here, we use a deep convolutional neural network to artificially enhance the resolution of NEXRAD PPI scans. The model is trained on 6 months of reflectivity observations from the Langley Hill, Washington, radar (KLGX), and we find that it substantially outperforms common interpolation schemes for 4× and 8× resolution increases based on several objective error and perceptual quality metrics.


2021 ◽  
Author(s):  
Biao Tong ◽  
Xiangfei Sun ◽  
Jiyang Fu ◽  
Yuncheng He ◽  
Pakwai Chan

Abstract. Tropical Cyclones (TCs) are one of the most destructive natural disasters. For the prevention and mitigation of TC-induced disasters, real-time monitoring and prediction of TCs is essential. At present, satellite cloud images (SCIs) are utilized widely as a basic data source for such studies. Although great achievements have been made in this field, lack of concerns on the identification of TC fingerprint from SCIs have become a potential issue, since it is a prerequisite step for follow-up analyses. This paper presents a methodology which identifies TC fingerprint via Deep Convolutional Neural Network (DCNN) techniques based on SCIs of more than 200 TCs over the Northwest Pacific basin. Two DCNN models have been proposed and validated, which are able to identify the TCs from not only single-TC featured SCIs but also multi-TCs featured SCIs. Results show that both models can reach 96 % of identification accuracy. As the TC intensity strengthens, the accuracy becomes better. To explore how these models work, heat maps are further extracted and analyzed. Results show that all the fingerprint features are focused on clouds during the testing process. For the majority of TC images, the cloud features in TC’s main parts, i.e., eye, eyewall and primary rainbands, are most emphasized, reflecting a consistent pattern with the subjective method.


2021 ◽  
Vol 2021 ◽  
pp. 1-12
Author(s):  
Huiling Lu

Based on the better generalization ability and the feature learning ability of the deep convolutional neural network, it is very significant to use the DCNN on the computer-aided diagnosis of a lung tumor. Firstly, a deep convolutional neural network was constructed according to the fuzzy characteristics and the complexity of lung CT images. Secondly, the relation between model parameters (iterations, different resolution) and recognition rate is discussed. Thirdly, the effects of different model structures for the identification of a lung tumor were analyzed by changing convolution kernel size, feature dimension, and depth of the network. Fourthly, the different optimization methods on how to influence the DCNN performance were discussed from three aspects containing pooling methods (maximum pooling and mean pooling), activation function (sigmoid and ReLU), and training algorithm (batch gradient descent and gradient descent with momentum). Finally, the experimental results verified the feasibility of DCNN used on computer-aided diagnosis of lung tumors, and it can achieve a good recognition rate when selecting the appropriate model parameters and model structure and using the method of gradient descent with momentum.


Complexity ◽  
2020 ◽  
Vol 2020 ◽  
pp. 1-11 ◽  
Author(s):  
Zhijian Huang ◽  
Bowen Sui ◽  
Jiayi Wen ◽  
Guohe Jiang

The shipping industry is developing towards intelligence rapidly. An accurate and fast method for ship image/video detection and classification is of great significance for not only the port management, but also the safe driving of Unmanned Surface Vehicle (USV). Thus, this paper makes a self-built dataset for the ship image/video detection and classification, and its method based on an improved regressive deep convolutional neural network is presented. This method promotes the regressive convolutional neural network from four aspects. First, the feature extraction layer is lightweighted by referring to YOLOv2. Second, a new feature pyramid network layer is designed by improving its structure in YOLOv3. Third, a proper frame and scale suitable for ships are designed with a clustering algorithm to reduced 60% anchors. Last, the activation function is verified and optimized. Then, the detecting experiment on 7 types of ships shows that the proposed method has advantage compared with the YOLO series networks and other intelligent methods. This method can solve the problem of low recognition rate and real-time performance for ship image/video detection and classification with a small dataset. On the testing-set, the final mAP is 0.9209, the Recall is 0.9818, the AIOU is 0.7991, and the FPS is 78–80 in video detection. Thus, this method provides a highly accurate and real-time ship detection method for the intelligent port management and visual processing of the USV. In addition, the proposed regressive deep convolutional network also has a better comprehensive performance than that of YOLOv2/v3.


2019 ◽  
Vol 2019 ◽  
pp. 1-14 ◽  
Author(s):  
Yung-Hui Li ◽  
Nai-Ning Yeh ◽  
Shih-Jen Chen ◽  
Yu-Chien Chung

Diabetic retinopathy (DR) is a complication of long-standing diabetes, which is hard to detect in its early stage because it only shows a few symptoms. Nowadays, the diagnosis of DR usually requires taking digital fundus images, as well as images using optical coherence tomography (OCT). Since OCT equipment is very expensive, it will benefit both the patients and the ophthalmologists if an accurate diagnosis can be made, based solely on reading digital fundus images. In the paper, we present a novel algorithm based on deep convolutional neural network (DCNN). Unlike the traditional DCNN approach, we replace the commonly used max-pooling layers with fractional max-pooling. Two of these DCNNs with a different number of layers are trained to derive more discriminative features for classification. After combining features from metadata of the image and DCNNs, we train a support vector machine (SVM) classifier to learn the underlying boundary of distributions of each class. For the experiments, we used the publicly available DR detection database provided by Kaggle. We used 34,124 training images and 1,000 validation images to build our model and tested with 53,572 testing images. The proposed DR classifier classifies the stages of DR into five categories, labeled with an integer ranging between zero and four. The experimental results show that the proposed method can achieve a recognition rate up to 86.17%, which is higher than previously reported in the literature. In addition to designing a machine learning algorithm, we also develop an app called “Deep Retina.” Equipped with a handheld ophthalmoscope, the average person can take fundus images by themselves and obtain an immediate result, calculated by our algorithm. It is beneficial for home care, remote medical care, and self-examination.


Sign in / Sign up

Export Citation Format

Share Document