scholarly journals Data augmentation and shadow image classification for shadow detection

2021 ◽  
Author(s):  
Guoquan Li ◽  
Lingyun Wen ◽  
Zhengwen Huang ◽  
Ruiyang Xia ◽  
Yu Pang
Information ◽  
2021 ◽  
Vol 12 (6) ◽  
pp. 249
Author(s):  
Xin Jin ◽  
Yuanwen Zou ◽  
Zhongbing Huang

The cell cycle is an important process in cellular life. In recent years, some image processing methods have been developed to determine the cell cycle stages of individual cells. However, in most of these methods, cells have to be segmented, and their features need to be extracted. During feature extraction, some important information may be lost, resulting in lower classification accuracy. Thus, we used a deep learning method to retain all cell features. In order to solve the problems surrounding insufficient numbers of original images and the imbalanced distribution of original images, we used the Wasserstein generative adversarial network-gradient penalty (WGAN-GP) for data augmentation. At the same time, a residual network (ResNet) was used for image classification. ResNet is one of the most used deep learning classification networks. The classification accuracy of cell cycle images was achieved more effectively with our method, reaching 83.88%. Compared with an accuracy of 79.40% in previous experiments, our accuracy increased by 4.48%. Another dataset was used to verify the effect of our model and, compared with the accuracy from previous results, our accuracy increased by 12.52%. The results showed that our new cell cycle image classification system based on WGAN-GP and ResNet is useful for the classification of imbalanced images. Moreover, our method could potentially solve the low classification accuracy in biomedical images caused by insufficient numbers of original images and the imbalanced distribution of original images.


2021 ◽  
Vol 11 (15) ◽  
pp. 6721
Author(s):  
Jinyeong Wang ◽  
Sanghwan Lee

In increasing manufacturing productivity with automated surface inspection in smart factories, the demand for machine vision is rising. Recently, convolutional neural networks (CNNs) have demonstrated outstanding performance and solved many problems in the field of computer vision. With that, many machine vision systems adopt CNNs to surface defect inspection. In this study, we developed an effective data augmentation method for grayscale images in CNN-based machine vision with mono cameras. Our method can apply to grayscale industrial images, and we demonstrated outstanding performance in the image classification and the object detection tasks. The main contributions of this study are as follows: (1) We propose a data augmentation method that can be performed when training CNNs with industrial images taken with mono cameras. (2) We demonstrate that image classification or object detection performance is better when training with the industrial image data augmented by the proposed method. Through the proposed method, many machine-vision-related problems using mono cameras can be effectively solved by using CNNs.


Mathematics ◽  
2021 ◽  
Vol 9 (6) ◽  
pp. 624
Author(s):  
Stefan Rohrmanstorfer ◽  
Mikhail Komarov ◽  
Felix Mödritscher

With the always increasing amount of image data, it has become a necessity to automatically look for and process information in these images. As fashion is captured in images, the fashion sector provides the perfect foundation to be supported by the integration of a service or application that is built on an image classification model. In this article, the state of the art for image classification is analyzed and discussed. Based on the elaborated knowledge, four different approaches will be implemented to successfully extract features out of fashion data. For this purpose, a human-worn fashion dataset with 2567 images was created, but it was significantly enlarged by the performed image operations. The results show that convolutional neural networks are the undisputed standard for classifying images, and that TensorFlow is the best library to build them. Moreover, through the introduction of dropout layers, data augmentation and transfer learning, model overfitting was successfully prevented, and it was possible to incrementally improve the validation accuracy of the created dataset from an initial 69% to a final validation accuracy of 84%. More distinct apparel like trousers, shoes and hats were better classified than other upper body clothes.


Author(s):  
Tongtong Li ◽  
Qiang Lin ◽  
Yanru Guo ◽  
Shaofang Zhao ◽  
Xianwu Zeng ◽  
...  

Abstract Bone scan is widely used for surveying bone metastases caused by various solid tumors. Scintigraphic images are characterized by inferior spatial resolution, bringing a significant challenge to manual analysis of images by nuclear medicine physicians. We present in this work a new framework for automatically classifying scintigraphic images collected from patients clinically diagnosed with lung cancer. The framework consists of data preparation and image classification. In the data preparation stage, data augmentation is used to enlarge the dataset, followed by image fusion and thoracic region extraction. In the image classification stage, we use a self-defined convolutional neural network consisting of feature extraction, feature aggregation, and feature classification sub-networks. The developed multi-class classification network can not only predict whether a bone scan image contains bone metastasis but also tell which subcategory of lung cancer that a bone metastasis metastasized from is present in the image. Experimental evaluations on a set of clinical bone scan images have shown that the proposed multi-class classification network is workable for automated classification of metastatic images, with achieving average scores of 0.7392, 0.7592, 0.7242, and 0.7292 for accuracy, precision, recall, and F-1 score, respectively.


Author(s):  
Subramani Palanisamy Harisubramanyabalaji ◽  
Shafiq ur Réhman ◽  
Mattias Nyberg ◽  
Joakim Gustavsson

Author(s):  
Xiaofeng Liu ◽  
Yang Zou ◽  
Lingsheng Kong ◽  
Zhihui Diao ◽  
Junliang Yan ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document