scholarly journals A Transfer Learning Strategy for Owl Sound Classification by Using Image Classification Model with Audio Spectrogram

Author(s):  
Kevin William Gunawan ◽  
◽  
Alam Ahmad Hidayat ◽  
Tjeng Wawan Cenggoro ◽  
Bens Pardamean ◽  
...  
Author(s):  
Asmaa Abbas ◽  
Mohammed M. Abdelsamea ◽  
Mohamed Medhat Gaber

ABSTRACTDue to the high availability of large-scale annotated image datasets, knowledge transfer from pre-trained models showed outstanding performance in medical image classification. However, building a robust image classification model for datasets with data irregularity or imbalanced classes can be a very challenging task, especially in the medical imaging domain. In this paper, we propose a novel deep convolutional neural network, we called Self Supervised Super Sample Decomposition for Transfer learning (4S-DT) model. 4S-DT encourages a coarse-to-fine transfer learning from large-scale image recognition tasks to a specific chest X-ray image classification task using a generic self-supervised sample decomposition approach. Our main contribution is a novel self-supervised learning mechanism guided by a super sample decomposition of unlabelled chest X-ray images. 4S-DT helps in improving the robustness of knowledge transformation via a downstream learning strategy with a class-decomposition layer to simplify the local structure of the data. 4S-DT can deal with any irregularities in the image dataset by investigating its class boundaries using a downstream class-decomposition mechanism. We used 50,000 unlabelled chest X-ray images to achieve our coarse-to-fine transfer learning with an application to COVID-19 detection, as an exemplar. 4S-DT has achieved an accuracy of 97.54% in the detection of COVID-19 cases on an extended test set enriched by augmented images, out of which all real COVID-19 cases were detected, which was the highest accuracy obtained when compared to other methods.


2021 ◽  
Vol 1 (1) ◽  
Author(s):  
Abdul Jalil Rozaqi ◽  
Muhammad Rudyanto Arief ◽  
Andi Sunyoto

Potatoes are a plant that has many benefits for human life. The potato plant has a problem, namely a disease that attacks the leaves. Disease on potato leaves that is often encountered is early blight and late blight. Image processing is a method that can be used to assist farmers in identifying potato leaf disease by utilizing leaf images. Image processing method development has been done a lot, one of which is by using the Convolutional Neural Network (CNN) algorithm. The CNN method is a good image classification algorithm because its layer architecture can extract leaf image features in depth, however, determining a good CNN architectural model requires a lot of data. CNN architecture will become overfitting if it uses less data, where the classification model has high accuracy on training data but the accuracy becomes poor on test data or new data. This research utilizes the Transfer Learning method to avoid an overfit model when the data used is not ideal or too little. Transfer Learning is a method that uses the CNN architecture that has been trained by other data previously which is then used for image classification on the new data. The purpose of this research was to use the Transfer Learning method on CNN architecture to classify potato leaf images in identifying potato leaf disease. This research compares the Transfer Learning method used to find the best method. The results of the experiments in this research indicate that the Transfer Learning VGG-16 method has the best classification performance results, this method produces the highest accuracy value of 95%.


Author(s):  
Sweety Duseja

Abstract: Many algorithms have been developed as a result of recent advances in machine learning to handle a variety of challenges. In recent years, the most popular transfer learning method has allowed researchers and engineers to run experiments with minimal computing and time resources. To tackle the challenges of classification, product identification, product suggestion, and picture-based search, this research proposed a transfer learning strategy for Fashion image classification based on hybrid 2D-CNN pretrained by VGG-16 and AlexNet. Pre-processing, feature extraction, and classification are the three parts of the proposed system's implementation. We used the Fashion MNIST dataset, which consists of 50,000 fashion photos that have been classified. Training and validation datasets have been separated. In comparison to other conventional methodologies, the suggested transfer learning approach has higher training and validation accuracy and reduced loss. Keywords: Machine Learning, Transfer Learning, Convolutional Neural Network, Image Classification, VGG16, AlexNet, 2D CNN.


Author(s):  
Z. L. Cai ◽  
Q. Weng ◽  
S. Z. Ye

Abstract. With the deepening research and cross-fusion in the modern remote sensing image area, the classification of high spatial resolution remote sensing images has captured the attention of the researchers in the field of remote sensing. However, due to the serious phenomenon of “same object, different spectrum” and “same spectrum, different object” of high-resolution remote sensing image, the traditional classification strategy is hard to handle this challenge. In this paper, a remote sensing image scene classification model based on SENet and Inception-V3 is proposed by utilizing the deep learning method and transfer learning strategy. The model first adds a dropout layer before the full connection layer of the original Inception-V3 model to avoid over-fitting. Then we embed the SENet module into the Inception-V3 model for optimizing the network performance. In this paper, global average pooling is used as squeeze operation, and then two fully connected layers are used to construct a bottleneck structure. The model proposed in this paper is more non-linear, can better fit the complex correlation between channels, and greatly reduces the amount of parameters and computation. In the training process, this paper adopts the transfer learning strategy, makes full use of existing models and knowledge, improves training efficiency, and finally obtains scene classification results. The experimental results based on AID high-score remote sensing scene images show that SE-Inception has faster convergence speed and more stable training effect than the original Inception-V3 training. Compared with other traditional methods and deep learning networks, the improved model proposed in this paper has greater accuracy improvement.


2020 ◽  
Author(s):  
Xiaoyu He ◽  
Juan Su ◽  
Guangyu Wang ◽  
Kang Zhang ◽  
Navarini Alexander ◽  
...  

BACKGROUND Pemphigus vulgaris (PV) and bullous pemphigoid (BP) are two rare but severe inflammatory dermatoses. Due to the regional lack of trained dermatologists, many patients with these two diseases are misdiagnosed and therefore incorrectly treated. An artificial intelligence diagnosis framework would be highly adaptable for the early diagnosis of these two diseases. OBJECTIVE Design and evaluate an artificial intelligence diagnosis framework for PV and BP. METHODS The work was conducted on a dermatological dataset consisting of 17,735 clinical images and 346 patient metadata of bullous dermatoses. A two-stage diagnosis framework was designed, where the first stage trained a clinical image classification model to classify bullous dermatoses from five common skin diseases and normal skin and the second stage developed a multimodal classification model of clinical images and patient metadata to further differentiate PV and BP. RESULTS The clinical image classification model and the multimodal classification model achieved an area under the receiver operating characteristic curve (AUROC) of 0.998 and 0.942, respectively. On the independent test set of 20 PV and 20 BP cases, our multimodal classification model (sensitivity: 0.85, specificity: 0.95) performed better than the average of 27 junior dermatologists (sensitivity: 0.68, specificity: 0.78) and comparable to the average of 69 senior dermatologists (sensitivity: 0.80, specificity: 0.87). CONCLUSIONS Our diagnosis framework based on clinical images and patient metadata achieved expert-level identification of PV and BP, and is potential to be an effective tool for dermatologists in remote areas in the early diagnosis of these two diseases.


Sign in / Sign up

Export Citation Format

Share Document