Convolutional Neural Network‐based Object Detection Model to Identify Gastrointestinal Stromal Tumors in Endoscopic Ultrasound Images

Author(s):  
Chang Kyo Oh ◽  
Taewan Kim ◽  
Yu Kyung Cho ◽  
Dae Young Cheung ◽  
Bo‐In Lee ◽  
...  
2019 ◽  
Vol 11 (13) ◽  
pp. 1516 ◽  
Author(s):  
Chang Lai ◽  
Jiyao Xu ◽  
Jia Yue ◽  
Wei Yuan ◽  
Xiao Liu ◽  
...  

With the development of ground-based all-sky airglow imager (ASAI) technology, a large amount of airglow image data needs to be processed for studying atmospheric gravity waves. We developed a program to automatically extract gravity wave patterns in the ASAI images. The auto-extraction program includes a classification model based on convolutional neural network (CNN) and an object detection model based on faster region-based convolutional neural network (Faster R-CNN). The classification model selects the images of clear nights from all ASAI raw images. The object detection model locates the region of wave patterns. Then, the wave parameters (horizontal wavelength, period, direction, etc.) can be calculated within the region of the wave patterns. Besides auto-extraction, we applied a wavelength check to remove the interference of wavelike mist near the imager. To validate the auto-extraction program, a case study was conducted on the images captured in 2014 at Linqu (36.2°N, 118.7°E), China. Compared to the result of the manual check, the auto-extraction recognized less (28.9% of manual result) wave-containing images due to the strict threshold, but the result shows the same seasonal variation as the references. The auto-extraction program applies a uniform criterion to avoid the accidental error in manual distinction of gravity waves and offers a reliable method to process large ASAI images for efficiently studying the climatology of atmospheric gravity waves.


2021 ◽  
Vol 1 (100) ◽  
pp. 68-77
Author(s):  
YAROSLAV M. TROFIMENKO ◽  
EVGENIY V. ERSHOV

A model for detecting objects in the image and a model for identifying steel-teeming ladles are discussed in the article. The object detection model is based on the use of a convolutional neural network. The identification model is based on the comparison of steel-teeming ladle features specific to the production process. The authors describe the adaptations of models to the conditions of the YOLOv3 architecture and the parameters of steel-teeming ladles in steel production. Simulation results are given at the end of the article.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Lee Ming Jun Melvin ◽  
Rajesh Elara Mohan ◽  
Archana Semwal ◽  
Povendhan Palanisamy ◽  
Karthikeyan Elangovan ◽  
...  

AbstractDrain blockage is a crucial problem in the urban environment. It heavily affects the ecosystem and human health. Hence, routine drain inspection is essential for urban environment. Manual drain inspection is a tedious task and prone to accidents and water-borne diseases. This work presents a drain inspection framework using convolutional neural network (CNN) based object detection algorithm and in house developed reconfigurable teleoperated robot called ‘Raptor’. The CNN based object detection model was trained using a transfer learning scheme with our custom drain-blocking objects data-set. The efficiency of the trained CNN algorithm and drain inspection robot Raptor was evaluated through various real-time drain inspection field trial. The experimental results indicate that our trained object detection algorithm has detect and classified the drain blocking objects with 91.42% accuracy for both offline and online test images and is able to process 18 frames per second (FPS). Further, the maneuverability of the robot was evaluated from various open and closed drain environment. The field trial results ensure that the robot maneuverability was stable, and its mapping and localization is also accurate in a complex drain environment.


2020 ◽  
Vol 14 ◽  
Author(s):  
Guoqiang Chen ◽  
Bingxin Bai ◽  
Huailong Yi

Background: Background: The development of deep learning technology has promoted the industrial intelligence, and automatic driving vehicles have become a hot research direction. As to the problem that pavement potholes threaten the safety of automatic driving vehicles, the pothole detection under complex environment conditions is studied. Objective: The goal of the work is to propose a new model of pavement pothole detection based on convolutional neural network. The main contribution is that the Multi-level Feature Fusion Block and the Detector Cascading Block are designed and a series of detectors are cascaded together to improve the detection accuracy of the proposed model. Methods: A pothole detection model is designed based on the original object detection model. In the study, the Transfer Connection Block in the Object Detection Module is removed and the Multi-level Feature Fusion Block is redesigned. At the same time, a Detector Cascading Block with multi-step detection is designed. Detectors are connected directly to the feature map and cascaded. In addition, the structure skips the transformation step. Results: The proposed method can be used to detect potholes efficiently. The real-time and accuracy of the model are improved after adjusting the network parameters and redesigning the model structure. The maximum detection accuracy of the proposed model is 75.24%. Conclusion: The Multi-level Feature Fusion Block designed enhances the fusion of high and low layer feature information and is conducive to extracting a large amount of target information. The Detector Cascade Block is a detector with cascade structure, which can realize more accurate prediction of the object. In a word, the model designed has greatly improved the detection accuracy and speed, which lays a solid foundation for pavement pothole detection under complex environmental conditions.


2021 ◽  
Vol 11 (23) ◽  
pp. 11275
Author(s):  
Atsushi Teramoto ◽  
Tomoyuki Shibata ◽  
Hyuga Yamada ◽  
Yoshiki Hirooka ◽  
Kuniaki Saito ◽  
...  

Upper gastrointestinal endoscopy is widely performed to detect early gastric cancers. As an automated detection method for early gastric cancer from endoscopic images, a method involving an object detection model, which is a deep learning technique, was proposed. However, there were challenges regarding the reduction in false positives in the detected results. In this study, we proposed a novel object detection model, U-Net R-CNN, based on a semantic segmentation technique that extracts target objects by performing a local analysis of the images. U-Net was introduced as a semantic segmentation method to detect early candidates for gastric cancer. These candidates were classified as gastric cancer cases or false positives based on box classification using a convolutional neural network. In the experiments, the detection performance was evaluated via the 5-fold cross-validation method using 1208 images of healthy subjects and 533 images of gastric cancer patients. When DenseNet169 was used as the convolutional neural network for box classification, the detection sensitivity and the number of false positives evaluated on a lesion basis were 98% and 0.01 per image, respectively, which improved the detection performance compared to the previous method. These results indicate that the proposed method will be useful for the automated detection of early gastric cancer from endoscopic images.


2021 ◽  
Vol 18 (1) ◽  
pp. 172988142199332
Author(s):  
Xintao Ding ◽  
Boquan Li ◽  
Jinbao Wang

Indoor object detection is a very demanding and important task for robot applications. Object knowledge, such as two-dimensional (2D) shape and depth information, may be helpful for detection. In this article, we focus on region-based convolutional neural network (CNN) detector and propose a geometric property-based Faster R-CNN method (GP-Faster) for indoor object detection. GP-Faster incorporates geometric property in Faster R-CNN to improve the detection performance. In detail, we first use mesh grids that are the intersections of direct and inverse proportion functions to generate appropriate anchors for indoor objects. After the anchors are regressed to the regions of interest produced by a region proposal network (RPN-RoIs), we then use 2D geometric constraints to refine the RPN-RoIs, in which the 2D constraint of every classification is a convex hull region enclosing the width and height coordinates of the ground-truth boxes on the training set. Comparison experiments are implemented on two indoor datasets SUN2012 and NYUv2. Since the depth information is available in NYUv2, we involve depth constraints in GP-Faster and propose 3D geometric property-based Faster R-CNN (DGP-Faster) on NYUv2. The experimental results show that both GP-Faster and DGP-Faster increase the performance of the mean average precision.


Electronics ◽  
2021 ◽  
Vol 10 (14) ◽  
pp. 1737
Author(s):  
Wooseop Lee ◽  
Min-Hee Kang ◽  
Jaein Song ◽  
Keeyeon Hwang

As automated vehicles have been considered one of the important trends in intelligent transportation systems, various research is being conducted to enhance their safety. In particular, the importance of technologies for the design of preventive automated driving systems, such as detection of surrounding objects and estimation of distance between vehicles. Object detection is mainly performed through cameras and LiDAR, but due to the cost and limits of LiDAR’s recognition distance, the need to improve Camera recognition technique, which is relatively convenient for commercialization, is increasing. This study learned convolutional neural network (CNN)-based faster regions with CNN (Faster R-CNN) and You Only Look Once (YOLO) V2 to improve the recognition techniques of vehicle-mounted monocular cameras for the design of preventive automated driving systems, recognizing surrounding vehicles in black box highway driving videos and estimating distances from surrounding vehicles through more suitable models for automated driving systems. Moreover, we learned the PASCAL visual object classes (VOC) dataset for model comparison. Faster R-CNN showed similar accuracy, with a mean average precision (mAP) of 76.4 to YOLO with a mAP of 78.6, but with a Frame Per Second (FPS) of 5, showing slower processing speed than YOLO V2 with an FPS of 40, and a Faster R-CNN, which we had difficulty detecting. As a result, YOLO V2, which shows better performance in accuracy and processing speed, was determined to be a more suitable model for automated driving systems, further progressing in estimating the distance between vehicles. For distance estimation, we conducted coordinate value conversion through camera calibration and perspective transform, set the threshold to 0.7, and performed object detection and distance estimation, showing more than 80% accuracy for near-distance vehicles. Through this study, it is believed that it will be able to help prevent accidents in automated vehicles, and it is expected that additional research will provide various accident prevention alternatives such as calculating and securing appropriate safety distances, depending on the vehicle types.


Sign in / Sign up

Export Citation Format

Share Document