scholarly journals CXR-RefineDet: Single-Shot Refinement Neural Network for Chest X-Ray Radiograph Based on Multiple Lesions Detection

2022 ◽  
Vol 2022 ◽  
pp. 1-11
Author(s):  
Cong Lin ◽  
Yongbin Zheng ◽  
Xiuchun Xiao ◽  
Jialun Lin

The workload of radiologists has dramatically increased in the context of the COVID-19 pandemic, causing misdiagnosis and missed diagnosis of diseases. The use of artificial intelligence technology can assist doctors in locating and identifying lesions in medical images. In order to improve the accuracy of disease diagnosis in medical imaging, we propose a lung disease detection neural network that is superior to the current mainstream object detection model in this paper. By combining the advantages of RepVGG block and Resblock in information fusion and information extraction, we design a backbone RRNet with few parameters and strong feature extraction capabilities. After that, we propose a structure called Information Reuse, which can solve the problem of low utilization of the original network output features by connecting the normalized features back to the network. Combining the network of RRNet and the improved RefineDet, we propose the overall network which was called CXR-RefineDet. Through a large number of experiments on the largest public lung chest radiograph detection dataset VinDr-CXR, it is found that the detection accuracy and inference speed of CXR-RefineDet have reached 0.1686 mAP and 6.8 fps, respectively, which is better than the two-stage object detection algorithm using a strong backbone like ResNet-50 and ResNet-101. In addition, the fast reasoning speed of CXR-RefineDet also provides the possibility for the actual implementation of the computer-aided diagnosis system.

2021 ◽  
Vol 56 (2) ◽  
pp. 235-248
Author(s):  
Fatchul Arifin ◽  
Herjuna Artanto ◽  
Nurhasanah ◽  
Teddy Surya Gunawan

COVID-19 is a new disease with a very rapid and tremendous spread. The most important thing needed now is a COVID-19 early detection system that is fast, easy to use, portable, and affordable. Various studies on desktop-based detection using Convolutional Neural Networks have been successfully conducted. However, no research has yet applied mobile-based detection, which requires low computational cost. Therefore, this research aims to produce a COVID-19 early detection system based on chest X-ray images using Convolutional Neural Network models to be deployed in mobile applications. It is expected that the proposed Convolutional Neural Network models can detect COVID-19 quickly, economically, and accurately. The used architecture is MobileNet's Single Shot Detection. The advantage of the Single Shot Detection MobileNet models is that they are lightweight to be applied to mobile-based devices. Therefore, these two versions will also be tested, which one is better. Both models have successfully detected COVID-19, normal, and viral pneumonia conditions with an average overall accuracy of 93.24% based on the test results. The Single Shot Detection MobileNet V1 model can detect COVID-19 with an average accuracy of 83.7%, while the Single Shot Detection MobileNet V2 Single Shot Detection model can detect COVID-19 with an average accuracy of 87.5%. Based on the research conducted, it can be concluded that the approach to detecting chest X-rays of COVID-19 can be detected using the MobileNet Single Shot Detection model. Besides, the V2 model shows better performance than the V1. Therefore, this model can be applied to increase the speed and affordability of COVID-19 detection.


2021 ◽  
Vol 143 (7) ◽  
Author(s):  
Yanbiao Zou ◽  
Mingquan Zhu ◽  
Xiangzhi Chen

Abstract Accurate locating of the weld seam under strong noise is the biggest challenge for automated welding. In this paper, we construct a robust seam detector on the framework of deep learning object detection algorithm. The representative object algorithm, a single shot multibox detector (SSD), is studied to establish the seam detector framework. The improved SSD is applied to seam detection. Under the SSD object detection framework, combined with the characteristics of the seam detection task, the multifeature combination network (MFCN) is proposed. The network comprehensively utilizes the local information and global information carried by the multilayer features to detect a weld seam and realizes the rapid and accurate detection of the weld seam. To solve the problem of single-frame seam image detection algorithm failure under continuous super-strong noise, the sequence image multifeature combination network (SMFCN) is proposed based on the MFCN detector. The recurrent neural network (RNN) is used to learn the temporal context information of convolutional features to accurately detect the seam under continuous super-noise. Experimental results show that the proposed seam detectors are extremely robust. The SMFCN can maintain extremely high detection accuracy under continuous super-strong noise. The welding results show that the laser vision seam tracking system using the SMFCN can ensure that the welding precision meets industrial requirements under a welding current of 150 A.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Lee Ming Jun Melvin ◽  
Rajesh Elara Mohan ◽  
Archana Semwal ◽  
Povendhan Palanisamy ◽  
Karthikeyan Elangovan ◽  
...  

AbstractDrain blockage is a crucial problem in the urban environment. It heavily affects the ecosystem and human health. Hence, routine drain inspection is essential for urban environment. Manual drain inspection is a tedious task and prone to accidents and water-borne diseases. This work presents a drain inspection framework using convolutional neural network (CNN) based object detection algorithm and in house developed reconfigurable teleoperated robot called ‘Raptor’. The CNN based object detection model was trained using a transfer learning scheme with our custom drain-blocking objects data-set. The efficiency of the trained CNN algorithm and drain inspection robot Raptor was evaluated through various real-time drain inspection field trial. The experimental results indicate that our trained object detection algorithm has detect and classified the drain blocking objects with 91.42% accuracy for both offline and online test images and is able to process 18 frames per second (FPS). Further, the maneuverability of the robot was evaluated from various open and closed drain environment. The field trial results ensure that the robot maneuverability was stable, and its mapping and localization is also accurate in a complex drain environment.


2020 ◽  
Vol 14 ◽  
Author(s):  
Guoqiang Chen ◽  
Bingxin Bai ◽  
Huailong Yi

Background: Background: The development of deep learning technology has promoted the industrial intelligence, and automatic driving vehicles have become a hot research direction. As to the problem that pavement potholes threaten the safety of automatic driving vehicles, the pothole detection under complex environment conditions is studied. Objective: The goal of the work is to propose a new model of pavement pothole detection based on convolutional neural network. The main contribution is that the Multi-level Feature Fusion Block and the Detector Cascading Block are designed and a series of detectors are cascaded together to improve the detection accuracy of the proposed model. Methods: A pothole detection model is designed based on the original object detection model. In the study, the Transfer Connection Block in the Object Detection Module is removed and the Multi-level Feature Fusion Block is redesigned. At the same time, a Detector Cascading Block with multi-step detection is designed. Detectors are connected directly to the feature map and cascaded. In addition, the structure skips the transformation step. Results: The proposed method can be used to detect potholes efficiently. The real-time and accuracy of the model are improved after adjusting the network parameters and redesigning the model structure. The maximum detection accuracy of the proposed model is 75.24%. Conclusion: The Multi-level Feature Fusion Block designed enhances the fusion of high and low layer feature information and is conducive to extracting a large amount of target information. The Detector Cascade Block is a detector with cascade structure, which can realize more accurate prediction of the object. In a word, the model designed has greatly improved the detection accuracy and speed, which lays a solid foundation for pavement pothole detection under complex environmental conditions.


2018 ◽  
Vol 2018 ◽  
pp. 1-10 ◽  
Author(s):  
Guo X. Hu ◽  
Zhong Yang ◽  
Lei Hu ◽  
Li Huang ◽  
Jia M. Han

The existing object detection algorithm based on the deep convolution neural network needs to carry out multilevel convolution and pooling operations to the entire image in order to extract a deep semantic features of the image. The detection models can get better results for big object. However, those models fail to detect small objects that have low resolution and are greatly influenced by noise because the features after repeated convolution operations of existing models do not fully represent the essential characteristics of the small objects. In this paper, we can achieve good detection accuracy by extracting the features at different convolution levels of the object and using the multiscale features to detect small objects. For our detection model, we extract the features of the image from their third, fourth, and 5th convolutions, respectively, and then these three scales features are concatenated into a one-dimensional vector. The vector is used to classify objects by classifiers and locate position information of objects by regression of bounding box. Through testing, the detection accuracy of our model for small objects is 11% higher than the state-of-the-art models. In addition, we also used the model to detect aircraft in remote sensing images and achieved good results.


Author(s):  
Wei Qiang ◽  
Yuyao He ◽  
Yujin Guo ◽  
Baoqi Li ◽  
Lingjiao He

As the in-depth exploration of oceans continues, the accurate and rapid detection of fish, bionics and other intelligent bodies in an underwater environment is more and more important for improving an underwater defense system. Because of the low accuracy and poor real-time performance of target detection in the complex underwater environment, we propose a target detection algorithm based on the improved SSD. We use the ResNet convolution neural network instead of the VGG convolution neural network of the SSD as the basic network for target detection. In the basic network, the depthwise-separated deformable convolution module proposed in this paper is used to extract the features of an underwater target so as to improve the target detection accuracy and speed in the complex underwater environment. It mainly fuses the depthwise separable convolution when the deformable convolution acquires the offset of a convolution core, thus reducing the number of parameters and achieving the purposes of increasing the speed of the convolution neural network and enhancing its robustness through sparse representation. The experimental results show that, compared with the SSD detection model that uses the ResNet convolution neural network as the basic network, the improved SSD detection model that uses the depthwise-separated deformable convolution module improves the accuracy of underwater target detection by 11 percentage points and reduces the detection time by 3 ms, thus validating the effectiveness of the algorithm proposed in the paper.


2021 ◽  
Vol 11 (8) ◽  
pp. 3531
Author(s):  
Hesham M. Eraqi ◽  
Karim Soliman ◽  
Dalia Said ◽  
Omar R. Elezaby ◽  
Mohamed N. Moustafa ◽  
...  

Extensive research efforts have been devoted to identify and improve roadway features that impact safety. Maintaining roadway safety features relies on costly manual operations of regular road surveying and data analysis. This paper introduces an automatic roadway safety features detection approach, which harnesses the potential of artificial intelligence (AI) computer vision to make the process more efficient and less costly. Given a front-facing camera and a global positioning system (GPS) sensor, the proposed system automatically evaluates ten roadway safety features. The system is composed of an oriented (or rotated) object detection model, which solves an orientation encoding discontinuity problem to improve detection accuracy, and a rule-based roadway safety evaluation module. To train and validate the proposed model, a fully-annotated dataset for roadway safety features extraction was collected covering 473 km of roads. The proposed method baseline results are found encouraging when compared to the state-of-the-art models. Different oriented object detection strategies are presented and discussed, and the developed model resulted in improving the mean average precision (mAP) by 16.9% when compared with the literature. The roadway safety feature average prediction accuracy is 84.39% and ranges between 91.11% and 63.12%. The introduced model can pervasively enable/disable autonomous driving (AD) based on safety features of the road; and empower connected vehicles (CV) to send and receive estimated safety features, alerting drivers about black spots or relatively less-safe segments or roads.


2020 ◽  
Vol 13 (1) ◽  
pp. 23
Author(s):  
Wei Zhao ◽  
William Yamada ◽  
Tianxin Li ◽  
Matthew Digman ◽  
Troy Runge

In recent years, precision agriculture has been researched to increase crop production with less inputs, as a promising means to meet the growing demand of agriculture products. Computer vision-based crop detection with unmanned aerial vehicle (UAV)-acquired images is a critical tool for precision agriculture. However, object detection using deep learning algorithms rely on a significant amount of manually prelabeled training datasets as ground truths. Field object detection, such as bales, is especially difficult because of (1) long-period image acquisitions under different illumination conditions and seasons; (2) limited existing prelabeled data; and (3) few pretrained models and research as references. This work increases the bale detection accuracy based on limited data collection and labeling, by building an innovative algorithms pipeline. First, an object detection model is trained using 243 images captured with good illimitation conditions in fall from the crop lands. In addition, domain adaptation (DA), a kind of transfer learning, is applied for synthesizing the training data under diverse environmental conditions with automatic labels. Finally, the object detection model is optimized with the synthesized datasets. The case study shows the proposed method improves the bale detecting performance, including the recall, mean average precision (mAP), and F measure (F1 score), from averages of 0.59, 0.7, and 0.7 (the object detection) to averages of 0.93, 0.94, and 0.89 (the object detection + DA), respectively. This approach could be easily scaled to many other crop field objects and will significantly contribute to precision agriculture.


2021 ◽  
Vol 11 (13) ◽  
pp. 6016
Author(s):  
Jinsoo Kim ◽  
Jeongho Cho

For autonomous vehicles, it is critical to be aware of the driving environment to avoid collisions and drive safely. The recent evolution of convolutional neural networks has contributed significantly to accelerating the development of object detection techniques that enable autonomous vehicles to handle rapid changes in various driving environments. However, collisions in an autonomous driving environment can still occur due to undetected obstacles and various perception problems, particularly occlusion. Thus, we propose a robust object detection algorithm for environments in which objects are truncated or occluded by employing RGB image and light detection and ranging (LiDAR) bird’s eye view (BEV) representations. This structure combines independent detection results obtained in parallel through “you only look once” networks using an RGB image and a height map converted from the BEV representations of LiDAR’s point cloud data (PCD). The region proposal of an object is determined via non-maximum suppression, which suppresses the bounding boxes of adjacent regions. A performance evaluation of the proposed scheme was performed using the KITTI vision benchmark suite dataset. The results demonstrate the detection accuracy in the case of integration of PCD BEV representations is superior to when only an RGB camera is used. In addition, robustness is improved by significantly enhancing detection accuracy even when the target objects are partially occluded when viewed from the front, which demonstrates that the proposed algorithm outperforms the conventional RGB-based model.


Sign in / Sign up

Export Citation Format

Share Document