Receptive field cooccurrence histograms for object detection

Author(s):  
S. Ekvall ◽  
D. Kragic
2020 ◽  
Vol 40 (4) ◽  
pp. 0415001
Author(s):  
谢学立 Xie Xueli ◽  
李传祥 Li Chuanxiang ◽  
杨小冈 Yang Xiaogang ◽  
席建祥 Xi Jianxiang ◽  
陈彤 Chen Tong

Sensors ◽  
2021 ◽  
Vol 21 (4) ◽  
pp. 1066
Author(s):  
Peng Jia ◽  
Fuxiang Liu

At present, the one-stage detector based on the lightweight model can achieve real-time speed, but the detection performance is challenging. To enhance the discriminability and robustness of the model extraction features and improve the detector’s detection performance for small objects, we propose two modules in this work. First, we propose a receptive field enhancement method, referred to as adaptive receptive field fusion (ARFF). It enhances the model’s feature representation ability by adaptively learning the fusion weights of different receptive field branches in the receptive field module. Then, we propose an enhanced up-sampling (EU) module to reduce the information loss caused by up-sampling on the feature map. Finally, we assemble ARFF and EU modules on top of YOLO v3 to build a real-time, high-precision and lightweight object detection system referred to as the ARFF-EU network. We achieve a state-of-the-art speed and accuracy trade-off on both the Pascal VOC and MS COCO data sets, reporting 83.6% AP at 37.5 FPS and 42.5% AP at 33.7 FPS, respectively. The experimental results show that our proposed ARFF and EU modules improve the detection performance of the ARFF-EU network and achieve the development of advanced, very deep detectors while maintaining real-time speed.


2020 ◽  
Vol 405 ◽  
pp. 138-148
Author(s):  
Lin Jiao ◽  
Shengyu Zhang ◽  
Shifeng Dong ◽  
Hongqiang Wang

Author(s):  
Luis A. Contreras ◽  
Abel Pacheco-Ortega ◽  
Jose I. Figueroa ◽  
Walterio W. Mayol-Cuevas ◽  
Jesus Savage

2021 ◽  
Vol 13 (13) ◽  
pp. 2538
Author(s):  
Xiaowu Xiao ◽  
Bo Wang ◽  
Lingjuan Miao ◽  
Linhao Li ◽  
Zhiqiang Zhou ◽  
...  

Infrared and visible images (multi-sensor or multi-band images) have many complementary features which can effectively boost the performance of object detection. Recently, convolutional neural networks (CNNs) have seen frequent use to perform object detection in multi-band images. However, it is very difficult for CNNs to extract complementary features from infrared and visible images. In order to solve this problem, a difference maximum loss function is proposed in this paper. The loss function can guide the learning directions of two base CNNs and maximize the difference between features from the two base CNNs, so as to extract complementary and diverse features. In addition, we design a focused feature-enhancement module to make features in the shallow convolutional layer more significant. In this way, the detection performance of small objects can be effectively improved while not increasing the computational cost in the testing stage. Furthermore, since the actual receptive field is usually much smaller than the theoretical receptive field, the deep convolutional layer would not have sufficient semantic features for accurate detection of large objects. To overcome this drawback, a cascaded semantic extension module is added to the deep layer. Through simple multi-branch convolutional layers and dilated convolutions with different dilation rates, the cascaded semantic extension module can effectively enlarge the actual receptive field and increase the detection accuracy of large objects. We compare our detection network with five other state-of-the-art infrared and visible image object detection networks. Qualitative and quantitative experimental results prove the superiority of the proposed detection network.


Sign in / Sign up

Export Citation Format

Share Document