ACFIM: Adaptively Cyclic Feature Information-Interaction Model for Object Detection

2021 ◽  
pp. 379-391
Author(s):  
Chen Song ◽  
Xu Cheng ◽  
Lihua Liu ◽  
Daqiu Li
2021 ◽  
Vol 13 (19) ◽  
pp. 3871
Author(s):  
Xu Cheng ◽  
Lihua Liu ◽  
Chen Song

Object detection and segmentation have recently shown encouraging results toward image analysis and interpretation due to their promising applications in remote sensing image fusion field. Although numerous methods have been proposed, implementing effective and efficient object detection is still very challenging for now, especially for the limitation of single modal data. The use of a single modal data is not always enough to reach proper spectral and spatial resolutions. The rapid expansion in the number and the availability of multi-source data causes new challenges for their effective and efficient processing. In this paper, we propose an effective feature information–interaction visual attention model for multimodal data segmentation and enhancement, which utilizes channel information to weight self-attentive feature maps of different sources, completing extraction, fusion, and enhancement of global semantic features with local contextual information of the object. Additionally, we further propose an adaptively cyclic feature information–interaction model, which adopts branch prediction to decide the number of visual perceptions, accomplishing adaptive fusion of global semantic features and local fine-grained information. Numerous experiments on several benchmarks show that the proposed approach can achieve significant improvements over baseline model.


Telecom IT ◽  
2019 ◽  
Vol 7 (4) ◽  
pp. 50-58
Author(s):  
M. Buinevich ◽  
P. Kurta

Research subject. Information interaction of the user with the information system. Objective. Improving the efficiency of user interaction with the information system to solve the main problem by customizing its interface and work script. Core results. The proposed methodology of scientific research aimed at achieving the goal, and consisting of 3 steps. As a result of each of them, the following main scientific results are expected to be obtained: interaction model, interaction assessment method, interaction optimization method. Also, it is expected to obtain private scientific results: the classification of the disadvantages of interaction, the influence of its parameters on the final efficiency, the architecture of the interface and scenario optimization system. Main conclusions. The proposed research scheme is scientifically correct and allows you to conduct a fullfledged scientific research and achieve the goal of the work. As a result, a method and a software tool will be developed that will make it possible to adjust a specific interface and a scenario for its work according to its own performance criteria - potency, operativeness and resource efficiency; at the same time, the general logic of solving the problem by the information system will remain unchanged.


2019 ◽  
Vol 9 (9) ◽  
pp. 1829 ◽  
Author(s):  
Jie Jiang ◽  
Hui Xu ◽  
Shichang Zhang ◽  
Yujie Fang

This study proposes a multiheaded object detection algorithm referred to as MANet. The main purpose of the study is to integrate feature layers of different scales based on the attention mechanism and to enhance contextual connections. To achieve this, we first replaced the feed-forward base network of the single-shot detector with the ResNet–101 (inspired by the Deconvolutional Single-Shot Detector) and then applied linear interpolation and the attention mechanism. The information of the feature layers at different scales was fused to improve the accuracy of target detection. The primary contributions of this study are the propositions of (a) a fusion attention mechanism, and (b) a multiheaded attention fusion method. Our final MANet detector model effectively unifies the feature information among the feature layers at different scales, thus enabling it to detect objects with different sizes and with higher precision. We used the 512 × 512 input MANet (the backbone is ResNet–101) to obtain a mean accuracy of 82.7% based on the PASCAL visual object class 2007 test. These results demonstrated that our proposed method yielded better accuracy than those provided by the conventional Single-shot detector (SSD) and other advanced detectors.


2008 ◽  
Vol 42 (1) ◽  
pp. 57-71 ◽  
Author(s):  
Jana Kludas ◽  
Eric Bruno ◽  
Stéphane Marchand-Maillet

Entropy ◽  
2021 ◽  
Vol 23 (7) ◽  
pp. 910
Author(s):  
Lei Yang ◽  
Jianchen Luo ◽  
Xiaowei Song ◽  
Menglong Li ◽  
Pengwei Wen ◽  
...  

A robust vehicle speed measurement system based on feature information fusion for vehicle multi-characteristic detection is proposed in this paper. A vehicle multi-characteristic dataset is constructed. With this dataset, seven CNN-based modern object detection algorithms are trained for vehicle multi-characteristic detection. The FPN-based YOLOv4 is selected as the best vehicle multi-characteristic detection algorithm, which applies feature information fusion of different scales with both rich high-level semantic information and detailed low-level location information. The YOLOv4 algorithm is improved by combing with the attention mechanism, in which the residual module in YOLOv4 is replaced by the ECA channel attention module with cross channel interaction. An improved ECA-YOLOv4 object detection algorithm based on both feature information fusion and cross channel interaction is proposed, which improves the performance of YOLOv4 for vehicle multi-characteristic detection and reduces the model parameter size and FLOPs as well. A multi-characteristic fused speed measurement system based on license plate, logo, and light is designed accordingly. The system performance is verified by experiments. The experimental results show that the speed measurement error rate of the proposed system meets the requirement of the China national standard GB/T 21555-2007 in which the speed measurement error rate should be less than 6%. The proposed system can efficiently enhance the vehicle speed measurement accuracy and effectively improve the vehicle speed measurement robustness.


2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Kanghua Hui ◽  
Jin Wang ◽  
Huaiqing He ◽  
W. H. Ip

Recently, tremendous strides have been made in generic object detection when used to detect faces, and there are still some remaining challenges. In this paper, a novel method is proposed named multilevel single stage network for face detection (MSNFD). Three breakthroughs are made in this research. Firstly, multilevel network is introduced into face detection to improve the efficiency of anchoring faces. Secondly, enhanced feature module is adopted to allow more feature information to be collected. Finally, two-stage weight loss function is employed to balance network of different levels. Experimental results on the WIDER FACE and FDDB datasets confirm that MSNFD has competitive accuracy to the mainstream methods, while keeping real-time performance.


2019 ◽  
Vol 9 (14) ◽  
pp. 2785 ◽  
Author(s):  
Yun Jiang ◽  
Tingting Peng ◽  
Ning Tan

Single Shot MultiBox Detector (SSD) has achieved good results in object detection but there are problems such as insufficient understanding of context information and loss of features in deep layers. In order to alleviate these problems, we propose a single-shot object detection network Context Perception-SSD (CP-SSD). CP-SSD promotes the network’s understanding of context information by using context information scene perception modules, so as to capture context information for objects of different scales. Deep layer feature map used semantic activation module, through self-supervised learning to adjust the context feature information and channel interdependence, and enhance useful semantic information. CP-SSD was validated on benchmark dataset PASCAL VOC 2007. The experimental results show that, compared with SSD, the mean Average Precision (mAP) of the CP-SSD detection method reaches 77.8%, which is 0.6% higher than that of SSD, and the detection effect was significantly improved in images with difficult to distinguish the object from the background.


Sign in / Sign up

Export Citation Format

Share Document