Colonoscopy image classification using self-supervised visual feature learning

Author(s):  
Nguyen Chi Thanh

Colonoscopy image classification is an image classification task that predicts whether colonoscopy images contain polyps or not. It is an important task input for an automatic polyp detection system. Recently, deep neural networks have been widely used for colonoscopy image classification due to the automatic feature extraction with high accuracy. However, training these networks requires a large amount of manually annotated data, which is expensive to acquire and limited by the available resources of endoscopy specialists. We propose a novel method for training colonoscopy image classification networks by using self-supervised visual feature learning to overcome this challenge. We adapt image denoising as a pretext task for self-supervised visual feature learning from unlabeled colonoscopy image dataset, where noise is added to the image for input, and the original image serves as the label. We use an unlabeled colonoscopy image dataset containing 8,500 images collected from the PACS system of Hospital 103 to train the pretext network. The feature exactor of the pretext network trained in a self-supervised way is used for colonoscopy image classification. A small labeled dataset from the public colonoscopy image dataset Kvasir is used to fine-tune the classifier. Our experiments demonstrate that the proposed self-supervised learning method can achieve a high colonoscopy image classification accuracy better than the classifier trained from scratch, especially at a small training dataset. When a dataset with only annotated 200 images is used for training classifiers, the proposed method improves accuracy from 72,16% to 93,15% compared to the baseline classifier.

2021 ◽  
Vol 207 ◽  
pp. 103206
Author(s):  
Cheng Xie ◽  
Ting Zeng ◽  
Hongxin Xiang ◽  
Keqin Li ◽  
Yun Yang ◽  
...  

IEEE Access ◽  
2021 ◽  
pp. 1-1
Author(s):  
Seyyed Mohammad Reza Hashemi ◽  
Hamid Hassanpour ◽  
Ehsan Kozegar ◽  
Tao Tan

Author(s):  
Yuchen Luo ◽  
Yi Zhang ◽  
Ming Liu ◽  
Yihong Lai ◽  
Panpan Liu ◽  
...  

Abstract Background and aims Improving the rate of polyp detection is an important measure to prevent colorectal cancer (CRC). Real-time automatic polyp detection systems, through deep learning methods, can learn and perform specific endoscopic tasks previously performed by endoscopists. The purpose of this study was to explore whether a high-performance, real-time automatic polyp detection system could improve the polyp detection rate (PDR) in the actual clinical environment. Methods The selected patients underwent same-day, back-to-back colonoscopies in a random order, with either traditional colonoscopy or artificial intelligence (AI)-assisted colonoscopy performed first by different experienced endoscopists (> 3000 colonoscopies). The primary outcome was the PDR. It was registered with clinicaltrials.gov. (NCT047126265). Results In this study, we randomized 150 patients. The AI system significantly increased the PDR (34.0% vs 38.7%, p < 0.001). In addition, AI-assisted colonoscopy increased the detection of polyps smaller than 6 mm (69 vs 91, p < 0.001), but no difference was found with regard to larger lesions. Conclusions A real-time automatic polyp detection system can increase the PDR, primarily for diminutive polyps. However, a larger sample size is still needed in the follow-up study to further verify this conclusion. Trial Registration clinicaltrials.gov Identifier: NCT047126265


IEEE Access ◽  
2021 ◽  
Vol 9 ◽  
pp. 64346-64357
Author(s):  
Reina Ishikawa ◽  
Ryo Hachiuma ◽  
Hideo Saito

Sensors ◽  
2021 ◽  
Vol 21 (5) ◽  
pp. 1820
Author(s):  
Xiaotao Shao ◽  
Qing Wang ◽  
Wei Yang ◽  
Yun Chen ◽  
Yi Xie ◽  
...  

The existing pedestrian detection algorithms cannot effectively extract features of heavily occluded targets which results in lower detection accuracy. To solve the heavy occlusion in crowds, we propose a multi-scale feature pyramid network based on ResNet (MFPN) to enhance the features of occluded targets and improve the detection accuracy. MFPN includes two modules, namely double feature pyramid network (FPN) integrated with ResNet (DFR) and repulsion loss of minimum (RLM). We propose the double FPN which improves the architecture to further enhance the semantic information and contours of occluded pedestrians, and provide a new way for feature extraction of occluded targets. The features extracted by our network can be more separated and clearer, especially those heavily occluded pedestrians. Repulsion loss is introduced to improve the loss function which can keep predicted boxes away from the ground truths of the unrelated targets. Experiments carried out on the public CrowdHuman dataset, we obtain 90.96% AP which yields the best performance, 5.16% AP gains compared to the FPN-ResNet50 baseline. Compared with the state-of-the-art works, the performance of the pedestrian detection system has been boosted with our method.


Mathematics ◽  
2021 ◽  
Vol 9 (6) ◽  
pp. 624
Author(s):  
Stefan Rohrmanstorfer ◽  
Mikhail Komarov ◽  
Felix Mödritscher

With the always increasing amount of image data, it has become a necessity to automatically look for and process information in these images. As fashion is captured in images, the fashion sector provides the perfect foundation to be supported by the integration of a service or application that is built on an image classification model. In this article, the state of the art for image classification is analyzed and discussed. Based on the elaborated knowledge, four different approaches will be implemented to successfully extract features out of fashion data. For this purpose, a human-worn fashion dataset with 2567 images was created, but it was significantly enlarged by the performed image operations. The results show that convolutional neural networks are the undisputed standard for classifying images, and that TensorFlow is the best library to build them. Moreover, through the introduction of dropout layers, data augmentation and transfer learning, model overfitting was successfully prevented, and it was possible to incrementally improve the validation accuracy of the created dataset from an initial 69% to a final validation accuracy of 84%. More distinct apparel like trousers, shoes and hats were better classified than other upper body clothes.


2019 ◽  
Vol 10 ◽  
pp. 2182-2191 ◽  
Author(s):  
Tushar C Jagadale ◽  
Dhanya S Murali ◽  
Shi-Wei Chu

Nonlinear nanoplasmonics is a largely unexplored research area that paves the way for many exciting applications, such as nanolasers, nanoantennas, and nanomodulators. In the field of nonlinear nanoplasmonics, it is highly desirable to characterize the nonlinearity of the optical absorption and scattering of single nanostructures. Currently, the common method to quantify optical nonlinearity is the z-scan technique, which yields real and imaginary parts of the permittivity by moving a thin sample with a laser beam. However, z-scan typically works with thin films, and thus acquires nonlinear responses from ensembles of nanostructures, not from single ones. In this work, we present an x-scan technique that is based on a confocal laser scanning microscope equipped with forward and backward detectors. The two-channel detection offers the simultaneous quantification for the nonlinear behavior of scattering, absorption and total attenuation by a single nanostructure. At low excitation intensities, both scattering and absorption responses are linear, thus confirming the linearity of the detection system. At high excitation intensities, we found that the nonlinear response can be derived directly from the point spread function of the x-scan images. Exceptionally large nonlinearities of both scattering and absorption are unraveled simultaneously for the first time. The present study not only provides a novel method for characterizing nonlinearity of a single nanostructure, but also reports surprisingly large plasmonic nonlinearities.


Sign in / Sign up

Export Citation Format

Share Document