scholarly journals Research on Lightweight Infrared Pedestrian Detection Model Algorithm for Embedded Platform

2021 ◽  
Vol 2021 ◽  
pp. 1-7
Author(s):  
Zhaoli Wu ◽  
Xin Wang ◽  
Chao Chen

Due to the limitation of energy consumption and power consumption, the embedded platform cannot meet the real-time requirements of the far-infrared image pedestrian detection algorithm. To solve this problem, this paper proposes a new real-time infrared pedestrian detection algorithm (RepVGG-YOLOv4, Rep-YOLO), which uses RepVGG to reconstruct the YOLOv4 backbone network, reduces the amount of model parameters and calculations, and improves the speed of target detection; using space spatial pyramid pooling (SPP) obtains different receptive field information to improve the accuracy of model detection; using the channel pruning compression method reduces redundant parameters, model size, and computational complexity. The experimental results show that compared with the YOLOv4 target detection algorithm, the Rep-YOLO algorithm reduces the model volume by 90%, the floating-point calculation is reduced by 93.4%, the reasoning speed is increased by 4 times, and the model detection accuracy after compression reaches 93.25%.

2021 ◽  
Vol 2078 (1) ◽  
pp. 012033
Author(s):  
Yuhuan Li ◽  
Jie Wang ◽  
Baodai Shi

Abstract The detection speed of target detection algorithm depends on the performance of computer equipment. Aiming at the problems of slow detection speed and difficult trade-off between detection accuracy and detection speed when the target detection model is used in embedded devices, a lightweight target detection model based on the improved Tiny YOLO-V3 is proposed. Firstly, we analyze the time consumption of each layer structure in the convolutional neural network, and do a lot of experiments and tests. Then, we compress the time-consuming structure substantially. Secondly, we propose the segmentation and fusion module to improve the detection accuracy. Finally, we use the remote sensing data set of Wuhan University for experiments, and the experimental results show that compared with Tiny YOLO-V3, the detection speed is improved by 4 times, and the accuracy is improved by 2 percentage points.


Complexity ◽  
2021 ◽  
Vol 2021 ◽  
pp. 1-8
Author(s):  
Li Zhang ◽  
Shaoqiang Wang ◽  
Hongwei Sun ◽  
Yifan Wang

Aiming at the problem that the embedded platform cannot meet the real-time detection of multisource images, this paper proposes a lightweight target detection network MNYOLO (MobileNet-YOLOv4-tiny) suitable for embedded platforms using deep separable convolution instead of standard convolution to reduce the number of model parameters and calculations; at the same time, the visible light target detection model is used as the pretraining model of the infrared target detection model and the infrared target data set collected on the spot is fine-tuned to obtain the infrared target detection model. On this basis, a decision-level fusion detection model is obtained to realize the complementary information of infrared and visible light multiband information. The experimental results show that it has a more obvious advantage in detection accuracy than the single-band target detection model while the decision-level fusion target detection model meets the real-time requirements and also verifies the effectiveness of the above algorithm.


2021 ◽  
Vol 13 (21) ◽  
pp. 4377
Author(s):  
Long Sun ◽  
Jie Chen ◽  
Dazheng Feng ◽  
Mengdao Xing

Unmanned aerial vehicle (UAV) is one of the main means of information warfare, such as in battlefield cruises, reconnaissance, and military strikes. Rapid detection and accurate recognition of key targets in UAV images are the basis of subsequent military tasks. The UAV image has characteristics of high resolution and small target size, and in practical application, the detection speed is often required to be fast. Existing algorithms are not able to achieve an effective trade-off between detection accuracy and speed. Therefore, this paper proposes a parallel ensemble deep learning framework for unmanned aerial vehicle video multi-target detection, which is a global and local joint detection strategy. It combines a deep learning target detection algorithm with template matching to make full use of image information. It also integrates multi-process and multi-threading mechanisms to speed up processing. Experiments show that the system has high detection accuracy for targets with focal lengths varying from one to ten times. At the same time, the real-time and stable display of detection results is realized by aiming at the moving UAV video image.


2021 ◽  
Vol 13 (19) ◽  
pp. 3949
Author(s):  
Ying Shen ◽  
Jie Li ◽  
Wenfu Lin ◽  
Liqiong Chen ◽  
Feng Huang ◽  
...  

The spectral information contained in the hyperspectral images (HSI) distinguishes the intrinsic properties of a target from the background, which is widely used in remote sensing. However, the low imaging speed and high data redundancy caused by the high spectral resolution of imaging spectrometers limit their application in scenarios with the real-time requirement. In this work, we achieve the precise detection of camouflaged targets based on snapshot multispectral imaging technology and band selection methods in urban-related scenes. Specifically, the camouflaged target detection algorithm combines the constrained energy minimization (CEM) algorithm and the improved maximum between-class variance (OTSU) algorithm (t-OTSU), which is proposed to obtain the initial target detection results and adaptively segment the target region. Moreover, an object region extraction (ORE) algorithm is proposed to obtain a complete target contour that improves the target detection capability of multispectral images (MSI). The experimental results show that the proposed algorithm has the ability to detect different camouflaged targets by using only four bands. The detection accuracy is above 99%, and the false alarm rate is below 0.2%. The research achieves the effective detection of camouflaged targets and has the potential to provide a new means for real-time multispectral sensing in complex scenes.


2021 ◽  
Vol 2021 ◽  
pp. 1-11
Author(s):  
Rui Wang ◽  
Ziyue Wang ◽  
Zhengwei Xu ◽  
Chi Wang ◽  
Qiang Li ◽  
...  

Object detection is an important part of autonomous driving technology. To ensure the safe running of vehicles at high speed, real-time and accurate detection of all the objects on the road is required. How to balance the speed and accuracy of detection is a hot research topic in recent years. This paper puts forward a one-stage object detection algorithm based on YOLOv4, which improves the detection accuracy and supports real-time operation. The backbone of the algorithm doubles the stacking times of the last residual block of CSPDarkNet53. The neck of the algorithm replaces the SPP with the RFB structure, improves the PAN structure of the feature fusion module, adds the attention mechanism CBAM and CA structure to the backbone and neck structure, and finally reduces the overall width of the network to the original 3/4, so as to reduce the model parameters and improve the inference speed. Compared with YOLOv4, the algorithm in this paper improves the average accuracy on KITTI dataset by 2.06% and BDD dataset by 2.95%. When the detection accuracy is almost unchanged, the inference speed of this algorithm is increased by 9.14%, and it can detect in real time at a speed of more than 58.47 FPS.


2022 ◽  
Vol 355 ◽  
pp. 03020
Author(s):  
Yitong Mao

The real-time pedestrian detection algorithm requires the model to be lightweight and robust. At the same time, the pedestrian object detection problem has the characteristics of aerial view Angle shooting, object overlap and weak light, etc. In order to design a more robust real-time detection model in weak light and crowded scene, this paper based on YOLO, raised a more efficient convolutional network. The experimental results show that, compared with YOLOX Network, the improved YOLO Network has a better detection effect in the lack of light scene and dense crowd scene, has a 5.0% advantage over YOLOX-s for pedestrians AP index, and has a 44.2% advantage over YOLOX-s for fps index.


Sensors ◽  
2020 ◽  
Vol 20 (13) ◽  
pp. 3646
Author(s):  
Jingwei Cao ◽  
Chuanxue Song ◽  
Silun Peng ◽  
Shixin Song ◽  
Xu Zhang ◽  
...  

Pedestrian detection is an important aspect of the development of intelligent vehicles. To address problems in which traditional pedestrian detection is susceptible to environmental factors and are unable to meet the requirements of accuracy in real time, this study proposes a pedestrian detection algorithm for intelligent vehicles in complex scenarios. YOLOv3 is one of the deep learning-based object detection algorithms with good performance at present. In this article, the basic principle of YOLOv3 is elaborated and analyzed firstly to determine its limitations in pedestrian detection. Then, on the basis of the original YOLOv3 network model, many improvements are made, including modifying grid cell size, adopting improved k-means clustering algorithm, improving multi-scale bounding box prediction based on receptive field, and using Soft-NMS algorithm. Finally, based on INRIA person and PASCAL VOC 2012 datasets, pedestrian detection experiments are conducted to test the performance of the algorithm in various complex scenarios. The experimental results show that the mean Average Precision (mAP) value reaches 90.42%, and the average processing time of each frame is 9.6 ms. Compared with other detection algorithms, the proposed algorithm exhibits accuracy and real-time performance together, good robustness and anti-interference ability in complex scenarios, strong generalization ability, high network stability, and detection accuracy and detection speed have been markedly improved. Such improvements are significant in protecting the road safety of pedestrians and reducing traffic accidents, and are conducive to ensuring the steady development of the technological level of intelligent vehicle driving assistance.


Author(s):  
Wei Qiang ◽  
Yuyao He ◽  
Yujin Guo ◽  
Baoqi Li ◽  
Lingjiao He

As the in-depth exploration of oceans continues, the accurate and rapid detection of fish, bionics and other intelligent bodies in an underwater environment is more and more important for improving an underwater defense system. Because of the low accuracy and poor real-time performance of target detection in the complex underwater environment, we propose a target detection algorithm based on the improved SSD. We use the ResNet convolution neural network instead of the VGG convolution neural network of the SSD as the basic network for target detection. In the basic network, the depthwise-separated deformable convolution module proposed in this paper is used to extract the features of an underwater target so as to improve the target detection accuracy and speed in the complex underwater environment. It mainly fuses the depthwise separable convolution when the deformable convolution acquires the offset of a convolution core, thus reducing the number of parameters and achieving the purposes of increasing the speed of the convolution neural network and enhancing its robustness through sparse representation. The experimental results show that, compared with the SSD detection model that uses the ResNet convolution neural network as the basic network, the improved SSD detection model that uses the depthwise-separated deformable convolution module improves the accuracy of underwater target detection by 11 percentage points and reduces the detection time by 3 ms, thus validating the effectiveness of the algorithm proposed in the paper.


Author(s):  
Leijin Long ◽  
Feng He ◽  
Hongjiang Liu

AbstractIn order to monitor the high-level landslides frequently occurring in Jinsha River area of Southwest China, and protect the lives and property safety of people in mountainous areas, the data of satellite remote sensing images are combined with various factors inducing landslides and transformed into landslide influence factors, which provides data basis for the establishment of landslide detection model. Then, based on the deep belief networks (DBN) and convolutional neural network (CNN) algorithm, two landslide detection models DBN and convolutional neural-deep belief network (CDN) are established to monitor the high-level landslide in Jinsha River. The influence of the model parameters on the landslide detection results is analyzed, and the accuracy of DBN and CDN models in dealing with actual landslide problems is compared. The results show that when the number of neurons in the DBN is 100, the overall error is the minimum, and when the number of learning layers is 3, the classification error is the minimum. The detection accuracy of DBN and CDN is 97.56% and 97.63%, respectively, which indicates that both DBN and CDN models are feasible in dealing with landslides from remote sensing images. This exploration provides a reference for the study of high-level landslide disasters in Jinsha River.


Sign in / Sign up

Export Citation Format

Share Document