scholarly journals Animal Species Recognition System using Deep Learning

Author(s):  
Deepthi K

Animals watching is a common hobby but to identify their species requires the assistance of Animal books. To provide Animal watchers a handy tool to admire the beauty of Animals, we developed a deep learning platform to assist users in recognizing species of Animals endemic to using app named the Imagenet of Animals (IoA). Animal images were learned by a convolutional neural network (CNN) to localize prominent features in the images. First, we established and generated a bounded region of interest to the shapes and colors of the object granularities and subsequently balanced the distribution of Animals species. Then, a skip connection method was used to linearly combine the outputs of the previous and current layers to improve feature extraction. Finally, we applied the SoftMax function to obtain a probability distribution of Animals features. The learned parameters of Animals features were used to identify pictures uploaded by mobile users. The proposed CNN model with skip connections achieved higher accuracy of 99.00 % compared with the 93.98% from a CNN and 89.00% from the SVM for the training images. As for the test dataset, the average sensitivity, specificity, and accuracy were 93.79%, 96.11%, and 95.37%, respectively.

2022 ◽  
Vol 8 ◽  
Author(s):  
Danyan Li ◽  
Xiaowei Han ◽  
Jie Gao ◽  
Qing Zhang ◽  
Haibo Yang ◽  
...  

Background: Multiparametric magnetic resonance imaging (mpMRI) plays an important role in the diagnosis of prostate cancer (PCa) in the current clinical setting. However, the performance of mpMRI usually varies based on the experience of the radiologists at different levels; thus, the demand for MRI interpretation warrants further analysis. In this study, we developed a deep learning (DL) model to improve PCa diagnostic ability using mpMRI and whole-mount histopathology data.Methods: A total of 739 patients, including 466 with PCa and 273 without PCa, were enrolled from January 2017 to December 2019. The mpMRI (T2 weighted imaging, diffusion weighted imaging, and apparent diffusion coefficient sequences) data were randomly divided into training (n = 659) and validation datasets (n = 80). According to the whole-mount histopathology, a DL model, including independent segmentation and classification networks, was developed to extract the gland and PCa area for PCa diagnosis. The area under the curve (AUC) were used to evaluate the performance of the prostate classification networks. The proposed DL model was subsequently used in clinical practice (independent test dataset; n = 200), and the PCa detective/diagnostic performance between the DL model and different level radiologists was evaluated based on the sensitivity, specificity, precision, and accuracy.Results: The AUC of the prostate classification network was 0.871 in the validation dataset, and it reached 0.797 using the DL model in the test dataset. Furthermore, the sensitivity, specificity, precision, and accuracy of the DL model for diagnosing PCa in the test dataset were 0.710, 0.690, 0.696, and 0.700, respectively. For the junior radiologist without and with DL model assistance, these values were 0.590, 0.700, 0.663, and 0.645 versus 0.790, 0.720, 0.738, and 0.755, respectively. For the senior radiologist, the values were 0.690, 0.770, 0.750, and 0.730 vs. 0.810, 0.840, 0.835, and 0.825, respectively. The diagnosis made with DL model assistance for radiologists were significantly higher than those without assistance (P < 0.05).Conclusion: The diagnostic performance of DL model is higher than that of junior radiologists and can improve PCa diagnostic accuracy in both junior and senior radiologists.


Author(s):  
Gabriel Maicas ◽  
Mathew Leonardi ◽  
Jodie Avery ◽  
Catrina Panuccio ◽  
Gustavo Carneiro ◽  
...  

Objectives: Pouch of Douglas (POD) obliteration is a severe consequence of inflammation in the pelvis, often seen in patients with endometriosis. The sliding sign is a dynamic transvaginal ultrasound (TVS) test that can diagnose POD obliteration. We aimed to develop a deep learning (DL) model to automatically classify the state of the POD using recorded videos depicting the sliding sign test. Methods: Expert sonologists performed, interpreted, and recorded videos of consecutive patients from Sept 2018-Apr 2020. The sliding sign was classified as positive (i.e. normal) or negative (i.e. POD obliteration). A DL model based on a temporal residual network was prospectively trained with a dataset of TVS videos. The model was tested on an independent test set and its diagnostic accuracy including area under the receiver operating characteristic curve (AUC), accuracy, sensitivity, specificity, positive and negative predictive value (PPV/NPV)) was compared to the reference standard sonologist classification (positive or negative sliding sign). Results: A positive sliding sign was depicted in 646/749 (86.2%) videos, whereas 103/749 (13.8%) videos depicted a negative sliding sign. The dataset was split into training (414 videos), validation (139), and testing (196) maintaining similar positive/negative proportions. When applied to the test dataset using a threshold of 0.9, the model achieved: AUC 96.5% (95%CI,90.8-100.0%), an accuracy of 88.8% (95%CI,83.5-92.8%), sensitivity of 88.6% (95%CI,83.0-92.9%), specificity of 90.0% (95%CI,68.3-98.8%), a PPV of 98.7% (95%CI,95.4-99.7%), and an NPV of 47.7% (95%CI,36.8-58.2%). Conclusions: We have developed an accurate DL model for the prediction of the TVS-based sliding sign classification.


Author(s):  
Yi Chung ◽  
Chih-Ang Chou ◽  
Chih-Yang Li

Identifying plants is not only the job of professionals, but also useful or essential for the plant lover and the general public. Although deep learning approaches for plant recognition are promising, driven by the success of convolutional neural networks (CNN), their performances are still far from the requirements of an in-field scenario. First, we propose a central attention concept that helps focus on the target instead of backgrounds in the image for tree species recognition. It could prevent model training from confused vision by establishing a dual path CNN deep learning framework, in which the central attention model combined with the CNN model based on InceptionV3 were employed to automatically extract the features. These two models were then learned together with a shared classification layer. Experimental results assessed the effectiveness of our proposed approach which outperformed each uni-path alone, and existing methods in the whole plant recognition system. Additionally, we created our own tree image database where each photo contained a wealth of information on the entire tree instead of an individual plant organ. Lastly, we developed a prototype system of an online/offline available tree species identification working on a consumer mobile platform that can identify the tree species not only by image recognition, but also detection and classification in real-time remotely.


Author(s):  
Daiju Ueda ◽  
Akira Yamamoto ◽  
Shoichi Ehara ◽  
Shinichi Iwata ◽  
Koji Abo ◽  
...  

Abstract Aims We aimed to develop models to detect aortic stenosis (AS) from chest radiographs—one of the most basic imaging tests—with artificial intelligence. Methods and Results We used 10433 retrospectively collected digital chest radiographs from 5638 patients to train, validate, and test three deep learning models. Chest radiographs were collected from patients who had also undergone echocardiography at a single institution between July 2016 and May 2019. These were labelled from the corresponding echocardiography assessments as AS-positive or AS-negative. The radiographs were separated on a patient basis into training (8327 images from 4512 patients, mean age 65 ± [SD] 15 years), validation (1041 images from 563 patients, mean age 65 ± 14 years), and test (1065 images from 563 patients, mean age 65 ± 14 years) datasets. The soft voting-based ensemble of the three developed models had the best overall performance for predicting AS with an AUC, sensitivity, specificity, accuracy, positive predictive value, and negative predictive value of 0.83 (95% CI 0.77–0.88), 0.78 (0.67–0.86), 0.71 (0.68–0.73), 0.71 (0.68–0.74), 0.18 (0.14–0.23), and 0.97 (0.96–0.98), respectively, in the validation dataset and 0.83 (0.78–0.88), 0.83 (0.74–0.90), 0.69 (0.66–0.72), 0.71 (0.68–0.73), 0.23 (0.19–0.28), and 0.97 (0.96–0.98), respectively, in the test dataset. Conclusion Deep learning models using chest radiographs have the potential to differentiate between radiographs of patients with and without AS. Lay summary We created AI models using deep learning to identify aortic stenosis from chest radiographs. Three AI models were developed and evaluated with 10433 retrospectively collected radiographs and labelled from echocardiography reports. The ensemble AI model could detect aortic stenosis in a test dataset with an AUC of 0.83 (95% CI 0.78–0.88). Since chest radiography is a cost effective and widely available imaging test, our model can provide an additive resource for the detection of aortic stenosis.


Author(s):  
Hajara Aliyu Abdulkarim ◽  
Mohd Azhar Abdul Razak ◽  
Rubita Sudirman ◽  
Norhafizah Ramli

Sickle cell anemia (SCA) is a serious hematological disorder, where affected patients are frequently hospitalized throughout a lifetime and even can cause death. The manual method of detecting and classifying abnormal cells of SCA patient blood film through a microscope is time-consuming, tedious, prone to error, and require a trained hematologist. The affected patient has many cell shapes that show important biomechanical characteristics. Hence, having an effective way of classifying the abnormalities present in the SCA disease will give a better insight into managing the concerned patient's life. This work proposed algorithm in two-phase firstly, automation of red blood cells (RBCs) extraction to identify the RBC region of interest (ROI) from the patient’s blood smear image. Secondly, deep learning AlexNet model is employed to classify and predict the abnormalities presence in SCA patients. The study was performed with (over 9,000 single RBC images) taken from 130 SCA patient each class having 750 cells. To develop a shape factor quantification and general multiscale shape analysis. We reveal that the proposed framework can classify 15 types of RBC shapes including normal in an automated manner with a deep AlexNet transfer learning model. The cell's name classification prediction accuracy, sensitivity, specificity, and precision of 95.92%, 77%, 98.82%, and 90% were achieved, respectively.


2020 ◽  
Vol 10 (7) ◽  
pp. 2547 ◽  
Author(s):  
Yu Song ◽  
Xu Qiao ◽  
Yutaro Iwamoto ◽  
Yen-wei Chen

Accurate automatic quantitative cephalometry are essential for orthodontics. However, manual labeling of cephalometric landmarks is tedious and subjective, which also must be performed by professional doctors. In recent years, deep learning has gained attention for its success in computer vision field. It has achieved large progress in resolving problems like image classification or image segmentation. In this paper, we propose a two-step method which can automatically detect cephalometric landmarks on skeletal X-ray images. First, we roughly extract a region of interest (ROI) patch for each landmark by registering the testing image to training images, which have annotated landmarks. Then, we utilize pre-trained networks with a backbone of ResNet50, which is a state-of-the-art convolutional neural network, to detect each landmark in each ROI patch. The network directly outputs the coordinates of the landmarks. We evaluate our method on two datasets: ISBI 2015 Grand Challenge in Dental X-ray Image Analysis and our own dataset provided by Shandong University. The experiments demonstrate that the proposed method can achieve satisfying results on both SDR (Successful Detection Rate) and SCR (Successful Classification Rate). However, the computational time issue remains to be improved in the future.


2021 ◽  
Author(s):  
Amishi Vijay ◽  
Jasleen Saini ◽  
B.S. Saini

A significant analysis is routine for Brain Tumor patients and it depends on accurate segmentation of Region of Interest. In automatic segmentation, field deep learning algorithms are attaining interest after they have performed very well in various ImageNet competitions. This review focuses on state-of-the-art Deep Learning Algorithms which are applied to Brain Tumor Segmentation. First, we review the methods of brain tumor segmentation, next the different deep learning algorithms and their performance measures like sensitivity, specificity and Dice similarity Coefficient (DSC) are discussed and Finally, we discuss and summarize the current deep learning techniques and identify future scope and trends.


Author(s):  
Lery Sakti Ramba

The purpose of this research is to design home automation system that can be controlled using voice commands. This research was conducted by studying other research related to the topics in this research, discussing with competent parties, designing systems, testing systems, and conducting analyzes based on tests that have been done. In this research voice recognition system was designed using Deep Learning Convolutional Neural Networks (DL-CNN). The CNN model that has been designed will then be trained to recognize several kinds of voice commands. The result of this research is a speech recognition system that can be used to control several electronic devices connected to the system. The speech recognition system in this research has a 100% success rate in room conditions with background intensity of 24dB (silent), 67.67% in room conditions with 42dB background noise intensity, and only 51.67% in room conditions with background intensity noise 52dB (noisy). The percentage of the success of the speech recognition system in this research is strongly influenced by the intensity of background noise in a room. Therefore, to obtain optimal results, the speech recognition system in this research is more suitable for use in rooms with low intensity background noise.


2020 ◽  
Vol 17 (3) ◽  
pp. 299-305 ◽  
Author(s):  
Riaz Ahmad ◽  
Saeeda Naz ◽  
Muhammad Afzal ◽  
Sheikh Rashid ◽  
Marcus Liwicki ◽  
...  

This paper presents a deep learning benchmark on a complex dataset known as KFUPM Handwritten Arabic TexT (KHATT). The KHATT data-set consists of complex patterns of handwritten Arabic text-lines. This paper contributes mainly in three aspects i.e., (1) pre-processing, (2) deep learning based approach, and (3) data-augmentation. The pre-processing step includes pruning of white extra spaces plus de-skewing the skewed text-lines. We deploy a deep learning approach based on Multi-Dimensional Long Short-Term Memory (MDLSTM) networks and Connectionist Temporal Classification (CTC). The MDLSTM has the advantage of scanning the Arabic text-lines in all directions (horizontal and vertical) to cover dots, diacritics, strokes and fine inflammation. The data-augmentation with a deep learning approach proves to achieve better and promising improvement in results by gaining 80.02% Character Recognition (CR) over 75.08% as baseline.


Sign in / Sign up

Export Citation Format

Share Document