scholarly journals Classification of the Trap-Neuter-Return Surgery Images of Stray Animals Using Yolo-Based Deep Learning Integrated with a Majority Voting System

2021 ◽  
Vol 11 (18) ◽  
pp. 8578
Author(s):  
Yi-Cheng Huang ◽  
Ting-Hsueh Chuang ◽  
Yeong-Lin Lai

Trap-neuter-return (TNR) has become an effective solution to reduce the prevalence of stray animals. Due to the non-culling policy for stray cats and dogs since 2017, there is a great demand for the sterilization of cats and dogs in Taiwan. In 2020, Heart of Taiwan Animal Care (HOTAC) had more than 32,000 cases of neutered cats and dogs. HOTAC needs to take pictures to record the ears and excised organs of each neutered cat or dog from different veterinary hospitals. The correctness of the archived medical photos and the different shooting and imaging angles from different veterinary hospitals must be carefully reviewed by human professionals. To reduce the cost of manual review, Yolo’s ensemble learning based on deep learning and a majority voting system can effectively identify TNR surgical images, save 80% of the labor force, and its average accuracy (mAP) exceeds 90%. The best feature extraction based on the Yolo model is Yolov4, whose mAP reaches 91.99%, and the result is integrated into the voting classification. Experimental results show that compared with the previous manual work, it can decrease the workload by more than 80%.

2021 ◽  
Author(s):  
Indrajeet Kumar ◽  
Jyoti Rawat

Abstract The manual diagnostic tests performed in laboratories for pandemic disease such as COVID19 is time-consuming, requires skills and expertise of the performer to yield accurate results. Moreover, it is very cost ineffective as the cost of test kits is high and also requires well-equipped labs to conduct them. Thus, other means of diagnosing the patients with presence of SARS-COV2 (the virus responsible for COVID19) must be explored. A radiography method like chest CT images is one such means that can be utilized for diagnosis of COVID19. The radio-graphical changes observed in CT images of COVID19 patient helps in developing a deep learning-based method for extraction of graphical features which are then used for automated diagnosis of the disease ahead of laboratory-based testing. The proposed work suggests an Artificial Intelligence (AI) based technique for rapid diagnosis of COVID19 from given volumetric CT images of patient’s chest by extracting its visual features and then using these features in the deep learning module. The proposed convolutional neural network is deployed for classifying the infectious and non-infectious SARS-COV2 subjects. The proposed network utilizes 746 chests scanned CT images of which 349 images belong to COVID19 positive cases while remaining 397 belong negative cases of COVID19. The extensive experiment has been completed with the accuracy of 98.4 %, sensitivity of 98.5 %, the specificity of 98.3 %, the precision of 97.1 %, F1score of 97.8 %. The obtained result shows the outstanding performance for classification of infectious and non-infectious for COVID19 cases.


10.29007/h46n ◽  
2022 ◽  
Author(s):  
Hoang Nhut Huynh ◽  
Minh Thanh Do ◽  
Gia Thinh Huynh ◽  
Anh Tu Tran ◽  
Trung Nghia Tran

Diabetic retinopathy (DR) is a complication of diabetes mellitus that causes retinal damage that can lead to vision loss if not detected and treated promptly. The common diagnosis stages of the disease take time, effort, and cost and can be misdiagnosed. In the recent period with the explosion of artificial intelligence, deep learning has become the most popular tool with high performance in many fields, especially in the analysis and classification of medical images. The Convolutional Neural Network (CNN) is more widely used as a deep learning method in medical imaging analysis with highly effective. In this paper, the five-stage image of modern DR (healthy, mild, moderate, severe, and proliferative) can be detected and classified using the deep learning technique. After cross-validation training and testing on the corresponding 5,590-image dataset, a pre-MobileNetV2 training model is proposed in classifying stages of diabetic retinopathy. The average accuracy of the model achieved was 93.89% with the precision of 94.00%, recall 92.00% and f1-score 90.00%. The corresponding thermal image is also given to help experts for evaluating the influence of the retina in each different stage.


2021 ◽  
Author(s):  
Pablo Millan Arias ◽  
Fatemeh Alipour ◽  
Kathleen Hill ◽  
Lila Kari

We present a novel Deep Learning method for the Unsupervised Classification of DNA Sequences (DeLUCS) that does not require sequence alignment, sequence homology, or (taxonomic) identifiers. DeLUCS uses Chaos Game Representations (CGRs) of primary DNA sequences, and generates “mimic” sequence CGRs to self-learn data patterns (genomic signatures) through the optimization of multiple neural networks. A majority voting scheme is then used to determine the final cluster label for each sequence. DeLUCS is able to cluster large and diverse datasets, with accuracies ranging from 77% to 100%: 2,500 complete vertebrate mitochondrial genomes, at taxonomic levels from sub-phylum to genera; 3,200 randomly selected 400 kbp-long bacterial genome segments, into families; three viral genome and gene datasets, averaging 1,300 sequences each, into virus subtypes. DeLUCS significantly outperforms two classic clustering methods (K-means and Gaussian Mixture Models) for unlabelled data, by as much as 48%. DeLUCS is highly effective, it is able to classify datasets of unlabelled primary DNA sequences totalling over 1 billion bp of data, and it bypasses common limitations to classification resulting from the lack of sequence homology, variation in sequence length, and the absence or instability of sequence annotations and taxonomic identifiers. Thus, DeLUCS offers fast and accurate DNA sequence classification for previously unclassifiable datasets.


Author(s):  
Hatem Keshk ◽  
Xu-Cheng Yin

Background: Deep Learning (DL) neural network methods have become a hotspot subject of research in the remote sensing field. Classification of aerial satellite images depends on spectral content, which is a challenging topic in remote sensing. Objective: With the aim to accomplish a high performance and accuracy of Egyptsat-1 satellite image classification, the use of the Convolutional Neural Network (CNN) is raised in this paper because CNN is considered a leading deep learning method. CNN is developed to classify aerial photographs into land cover classes such as urban, vegetation, desert, water bodies, soil, roads, etc. In our work, a comparison between MAXIMUM Likelihood (ML) which represents the traditional supervised classification methods and CNN method is conducted. Conclusion: This research finds that CNN outperforms ML by 9%. The convolutional neural network has better classification result, which reached 92.25% as its average accuracy. Also, the experiments showed that the convolutional neural network is the most satisfactory and effective classification method applied to classify Egyptsat-1 satellite images.


2021 ◽  
Vol 11 (5) ◽  
pp. 2149
Author(s):  
Moumita Sen Sarma ◽  
Kaushik Deb ◽  
Pranab Kumar Dhar ◽  
Takeshi Koshiba

Sports activities play a crucial role in preserving our health and mind. Due to the rapid growth of sports video repositories, automatized classification has become essential for easy access and retrieval, content-based recommendations, contextual advertising, etc. Traditional Bangladeshi sport is a genre of sports that bears the cultural significance of Bangladesh. Classification of this genre can act as a catalyst in reviving their lost dignity. In this paper, the Deep Learning method is utilized to classify traditional Bangladeshi sports videos by extracting both the spatial and temporal features from the videos. In this regard, a new Traditional Bangladeshi Sports Video (TBSV) dataset is constructed containing five classes: Boli Khela, Kabaddi, Lathi Khela, Kho Kho, and Nouka Baich. A key contribution of this paper is to develop a scratch model by incorporating the two most prominent deep learning algorithms: convolutional neural network (CNN) and long short term memory (LSTM). Moreover, the transfer learning approach with the fine-tuned VGG19 and LSTM is used for TBSV classification. Furthermore, the proposed model is assessed over four challenging datasets: KTH, UCF-11, UCF-101, and UCF Sports. This model outperforms some recent works on these datasets while showing 99% average accuracy on the TBSV dataset.


2019 ◽  
Vol 27 ◽  
pp. 03001
Author(s):  
Zhang Yue ◽  
Li Feng

The field of deep learning applications is becoming more widespread. The use of traditional algorithms for arrhythmia detection is cumbersome and the algorithm complexity is relatively high. Using the deep learning model to directly input data into the model will make it difficult to effectively segment the data, which will have a large error in the recognition accuracy. We collected data from six commonly used ECG databases and then used the label shuffling method to amplify the samples, effectively overcoming the effects of sample imbalance. The LSTM model is used for feature point detection to effectively locate the key feature points of the ECG signal, thereby completing the segmentation and processing of the data. Finally, the use of a multiple input model single output for arrhythmia detection has achieved significant results. The average accuracy of the final arrhythmia classification of the model reached 0.711. The accuracy of detecting the fusion of ventricular and normal beat and the aberrated atrial premature beat exceeds 0.9.


Agronomy ◽  
2020 ◽  
Vol 10 (11) ◽  
pp. 1721 ◽  
Author(s):  
Kunlong Yang ◽  
Weizhen Zhong ◽  
Fengguo Li

The segmentation and classification of leaves in plant images are a great challenge, especially when several leaves are overlapping in images with a complicated background. In this paper, the segmentation and classification of leaf images with a complicated background using deep learning are studied. First, more than 2500 leaf images with a complicated background are collected and artificially labeled with target pixels and background pixels. Two-thousand of them are fed into a Mask Region-based Convolutional Neural Network (Mask R-CNN) to train a model for leaf segmentation. Then, a training set that contains more than 1500 training images of 15 species is fed into a very deep convolutional network with 16 layers (VGG16) to train a model for leaf classification. The best hyperparameters for these methods are found by comparing a variety of parameter combinations. The results show that the average Misclassification Error (ME) of 80 test images using Mask R-CNN is 1.15%. The average accuracy value for the leaf classification of 150 test images using VGG16 is up to 91.5%. This indicates that these methods can be used to segment and classify the leaf image with a complicated background effectively. It could provide a reference for the phenotype analysis and automatic classification of plants.


Symmetry ◽  
2019 ◽  
Vol 12 (1) ◽  
pp. 8
Author(s):  
Jing Chen ◽  
Jun Feng ◽  
Xia Sun ◽  
Yang Liu

Sentiment classification of forum posts of massive open online courses is essential for educators to make interventions and for instructors to improve learning performance. Lacking monitoring on learners’ sentiments may lead to high dropout rates of courses. Recently, deep learning has emerged as an outstanding machine learning technique for sentiment classification, which extracts complex features automatically with rich representation capabilities. However, deep neural networks always rely on a large amount of labeled data for supervised training. Constructing large-scale labeled training datasets for sentiment classification is very laborious and time consuming. To address this problem, this paper proposes a co-training, semi-supervised deep learning model for sentiment classification, leveraging limited labeled data and massive unlabeled data simultaneously to achieve performance comparable to those methods trained on massive labeled data. To satisfy the condition of two views of co-training, we encoded texts into vectors from views of word embedding and character-based embedding independently, considering words’ external and internal information. To promote the classification performance with limited data, we propose a double-check strategy sample selection method to select samples with high confidence to augment the training set iteratively. In addition, we propose a mixed loss function both considering the labeled data with asymmetric and unlabeled data. Our proposed method achieved a 89.73% average accuracy and an 93.55% average F1-score, about 2.77% and 3.2% higher than baseline methods. Experimental results demonstrate the effectiveness of the proposed model trained on limited labeled data, which performs much better than those trained on massive labeled data.


2021 ◽  
Vol 11 (22) ◽  
pp. 10593
Author(s):  
Nabeela Kausar ◽  
Abdul Hameed ◽  
Mohsin Sattar ◽  
Ramiza Ashraf ◽  
Ali Shariq Imran ◽  
...  

Skin cancer is a widespread disease associated with eight diagnostic classes. The diagnosis of multiple types of skin cancer is a challenging task for dermatologists due to the similarity of skin cancer classes in phenotype. The average accuracy of multiclass skin cancer diagnosis is 62% to 80%. Therefore, the classification of skin cancer using machine learning can be beneficial in the diagnosis and treatment of the patients. Several researchers developed skin cancer classification models for binary class but could not extend the research to multiclass classification with better performance ratios. We have developed deep learning-based ensemble classification models for multiclass skin cancer classification. Experimental results proved that the individual deep learners perform better for skin cancer classification, but still the development of ensemble is a meaningful approach since it enhances the classification accuracy. Results show that the accuracy of individual learners of ResNet, InceptionV3, DenseNet, InceptionResNetV2, and VGG-19 are 72%, 91%, 91.4%, 91.7% and 91.8%, respectively. The accuracy of proposed majority voting and weighted majority voting ensemble models are 98% and 98.6%, respectively. The accuracy of proposed ensemble models is higher than the individual deep learners and the dermatologists’ diagnosis accuracy. The proposed ensemble models are compared with the recently developed skin cancer classification approaches. The results show that the proposed ensemble models outperform recently developed multiclass skin cancer classification models.


2022 ◽  
Vol 13 (2) ◽  
pp. 0-0

Pulmonary disease is widespread worldwide. There is persistent blockage of the lungs, pneumonia, asthma, TB, etc. It is essential to diagnose the lungs promptly. For this reason, machine learning models were developed. For lung disease prediction, many deep learning technologies, including the CNN, and the capsule network, are used. The fundamental CNN has low rotating, inclined, or other irregular image orientation efficiency. Therefore by integrating the space transformer network (STN) with CNN, we propose a new hybrid deep learning architecture named STNCNN. The new model is implemented on the dataset from the Kaggle repository for an NIH chest X-ray image. STNCNN has an accuracy of 69% in respect of the entire dataset, while the accuracy values of vanilla grey, vanilla RGB, hybrid CNN are 67.8%, 69.5%, and 63.8%, respectively. When the sample data set is applied, STNCNN takes much less time to train at the cost of a slightly less reliable validation. Therefore both specialists and physicians are simplified by the proposed STNCNN System for the diagnosis of lung disease.


Sign in / Sign up

Export Citation Format

Share Document