scholarly journals Object Detection Based on Fast/Faster RCNN Employing Fully Convolutional Architectures

2018 ◽  
Vol 2018 ◽  
pp. 1-7 ◽  
Author(s):  
Yun Ren ◽  
Changren Zhu ◽  
Shunping Xiao

Modern object detectors always include two major parts: a feature extractor and a feature classifier as same as traditional object detectors. The deeper and wider convolutional architectures are adopted as the feature extractor at present. However, many notable object detection systems such as Fast/Faster RCNN only consider simple fully connected layers as the feature classifier. In this paper, we declare that it is beneficial for the detection performance to elaboratively design deep convolutional networks (ConvNets) of various depths for feature classification, especially using the fully convolutional architectures. In addition, this paper also demonstrates how to employ the fully convolutional architectures in the Fast/Faster RCNN. Experimental results show that a classifier based on convolutional layer is more effective for object detection than that based on fully connected layer and that the better detection performance can be achieved by employing deeper ConvNets as the feature classifier.

Author(s):  
Pengxin Ding ◽  
Huan Zhou ◽  
Jinxia Shang ◽  
Xiang Zou ◽  
Minghui Wang

This paper designs a method that can generate anchors of various shapes for the object detection framework. This method has the characteristics of novelty and flexibility. Different from the previous anchors generated by a pre-defined manner, our anchors are generated dynamically by an anchor generator. Specially, the anchor generator is not fixed but learned from the hand-designed anchors, which means that our anchor generator is able to work well in various scenes. In the inference time, the weights of anchor generator are estimated by a simple network where the input is some hand-designed anchor. In addition, in order to make the difference between the number of positive and negative samples smaller, we use an adaptive IOU threshold related to the object size to solve this problem. At the same time, we proved that our proposed method is effective and conducted a lot of experiments on the COCO dataset. Experimental results show that after replacing the anchor generation method in the previous object detectors (such as SSD, mask RCNN, and Retinanet) with our proposed method, the detection performance of the model has been greatly improved compared to before the replacement, which proves our method is effective.


2021 ◽  
Vol 13 (9) ◽  
pp. 1854
Author(s):  
Syed Muhammad Arsalan Bashir ◽  
Yi Wang

This paper deals with detecting small objects in remote sensing images from satellites or any aerial vehicle by utilizing the concept of image super-resolution for image resolution enhancement using a deep-learning-based detection method. This paper provides a rationale for image super-resolution for small objects by improving the current super-resolution (SR) framework by incorporating a cyclic generative adversarial network (GAN) and residual feature aggregation (RFA) to improve detection performance. The novelty of the method is threefold: first, a framework is proposed, independent of the final object detector used in research, i.e., YOLOv3 could be replaced with Faster R-CNN or any object detector to perform object detection; second, a residual feature aggregation network was used in the generator, which significantly improved the detection performance as the RFA network detected complex features; and third, the whole network was transformed into a cyclic GAN. The image super-resolution cyclic GAN with RFA and YOLO as the detection network is termed as SRCGAN-RFA-YOLO, which is compared with the detection accuracies of other methods. Rigorous experiments on both satellite images and aerial images (ISPRS Potsdam, VAID, and Draper Satellite Image Chronology datasets) were performed, and the results showed that the detection performance increased by using super-resolution methods for spatial resolution enhancement; for an IoU of 0.10, AP of 0.7867 was achieved for a scale factor of 16.


Symmetry ◽  
2021 ◽  
Vol 13 (4) ◽  
pp. 678
Author(s):  
Vladimir Tadic ◽  
Tatjana Loncar-Turukalo ◽  
Akos Odry ◽  
Zeljen Trpovski ◽  
Attila Toth ◽  
...  

This note presents a fuzzy optimization of Gabor filter-based object and text detection. The derivation of a 2D Gabor filter and the guidelines for the fuzzification of the filter parameters are described. The fuzzy Gabor filter proved to be a robust text an object detection method in low-quality input images as extensively evaluated in the problem of license plate localization. The extended set of examples confirmed that the fuzzy optimized Gabor filter with adequately fuzzified parameters detected the desired license plate texture components and highly improved the object detection when compared to the classic Gabor filter. The robustness of the proposed approach was further demonstrated on other images of various origin containing text and different textures, captured using low-cost or modest quality acquisition procedures. The possibility to fine tune the fuzzification procedure to better suit certain applications offers the potential to further boost detection performance.


Author(s):  
Andrew Brock ◽  
Theodore Lim ◽  
J. M. Ritchie ◽  
Nick Weston

End-to-end machine analysis of engineering document drawings requires a reliable and precise vision frontend capable of localizing and classifying various characters in context. We develop an object detection framework, based on convolutional networks, designed specifically for optical character recognition in engineering drawings. Our approach enables classification and localization on a 10-fold cross-validation of an internal dataset for which other techniques prove unsuitable.


2021 ◽  
Vol 13 (22) ◽  
pp. 4517
Author(s):  
Falin Wu ◽  
Jiaqi He ◽  
Guopeng Zhou ◽  
Haolun Li ◽  
Yushuang Liu ◽  
...  

Object detection in remote sensing images plays an important role in both military and civilian remote sensing applications. Objects in remote sensing images are different from those in natural images. They have the characteristics of scale diversity, arbitrary directivity, and dense arrangement, which causes difficulties in object detection. For objects with a large aspect ratio and that are oblique and densely arranged, using an oriented bounding box can help to avoid deleting some correct detection bounding boxes by mistake. The classic rotational region convolutional neural network (R2CNN) has advantages for text detection. However, R2CNN has poor performance in the detection of slender objects with arbitrary directivity in remote sensing images, and its fault tolerance rate is low. In order to solve this problem, this paper proposes an improved R2CNN based on a double detection head structure and a three-point regression method, namely, TPR-R2CNN. The proposed network modifies the original R2CNN network structure by applying a double fully connected (2-fc) detection head and classification fusion. One detection head is for classification and horizontal bounding box regression, the other is for classification and oriented bounding box regression. The three-point regression method (TPR) is proposed for oriented bounding box regression, which determines the positions of the oriented bounding box by regressing the coordinates of the center point and the first two vertices. The proposed network was validated on the DOTA-v1.5 and HRSC2016 datasets, and it achieved a mean average precision (mAP) of 3.90% and 15.27%, respectively, from feature pyramid network (FPN) baselines with a ResNet-50 backbone.


Sensors ◽  
2018 ◽  
Vol 18 (9) ◽  
pp. 2929 ◽  
Author(s):  
Yuanyuan Wang ◽  
Chao Wang ◽  
Hong Zhang

With the capability to automatically learn discriminative features, deep learning has experienced great success in natural images but has rarely been explored for ship classification in high-resolution SAR images due to the training bottleneck caused by the small datasets. In this paper, convolutional neural networks (CNNs) are applied to ship classification by using SAR images with the small datasets. First, ship chips are constructed from high-resolution SAR images and split into training and validation datasets. Second, a ship classification model is constructed based on very deep convolutional networks (VGG). Then, VGG is pretrained via ImageNet, and fine tuning is utilized to train our model. Six scenes of COSMO-SkyMed images are used to evaluate our proposed model with regard to the classification accuracy. The experimental results reveal that (1) our proposed ship classification model trained by fine tuning achieves more than 95% average classification accuracy, even with 5-cross validation; (2) compared with other models, the ship classification model based on VGG16 achieves at least 2% higher accuracies for classification. These experimental results reveal the effectiveness of our proposed method.


Sign in / Sign up

Export Citation Format

Share Document