scholarly journals Learning Geometry Information of Target for Visual Object Tracking with Siamese Networks

Sensors ◽  
2021 ◽  
Vol 21 (23) ◽  
pp. 7790
Author(s):  
Hang Chen ◽  
Weiguo Zhang ◽  
Danghui Yan

Recently, Siamese architecture has been widely used in the field of visual tracking, and has achieved great success. Most Siamese network based trackers aggregate the target information of two branches by cross-correlation. However, since the location of the sampling points in the search feature area is pre-fixed in cross-correlation operation, these trackers suffer from either background noise influence or missing foreground information. Moreover, the cross-correlation between the template and the search area neglects the geometry information of the target. In this paper, we propose a Siamese deformable cross-correlation network to model the geometric structure of target and improve the performance of visual tracking. We propose to learn an offset field end-to-end in cross-correlation. With the guidance of the offset field, the sampling in the search image area can adapt to the deformation of the target, and realize the modeling of the geometric structure of the target. We further propose an online classification sub-network to model the variation of target appearance and enhance the robustness of the tracker. Extensive experiments are conducted on four challenging benchmarks, including OTB2015, VOT2018, VOT2019 and UAV123. The results demonstrate that our tracker achieves state-of-the-art performance.

Author(s):  
Zheng Zhu ◽  
Qiang Wang ◽  
Bo Li ◽  
Wei Wu ◽  
Junjie Yan ◽  
...  

2018 ◽  
Vol 77 (17) ◽  
pp. 22131-22143 ◽  
Author(s):  
Longchao Yang ◽  
Peilin Jiang ◽  
Fei Wang ◽  
Xuan Wang

2020 ◽  
Vol 34 (07) ◽  
pp. 13017-13024 ◽  
Author(s):  
Jinghao Zhou ◽  
Peng Wang ◽  
Haoyang Sun

The problem of visual object tracking has traditionally been handled by variant tracking paradigms, either learning a model of the object's appearance exclusively online or matching the object with the target in an offline-trained embedding space. Despite the recent success, each method agonizes over its intrinsic constraint. The online-only approaches suffer from a lack of generalization of the model they learn thus are inferior in target regression, while the offline-only approaches (e.g., convolutional siamese trackers) lack the target-specific context information thus are not discriminative enough to handle distractors, and robust enough to deformation. Therefore, we propose an online module with an attention mechanism for offline siamese networks to extract target-specific features under L2 error. We further propose a filter update strategy adaptive to treacherous background noises for discriminative learning, and a template update strategy to handle large target deformations for robust learning. Effectiveness can be validated in the consistent improvement over three siamese baselines: SiamFC, SiamRPN++, and SiamMask. Beyond that, our model based on SiamRPN++ obtains the best results over six popular tracking benchmarks and can operate beyond real-time.


2017 ◽  
Vol 2017 ◽  
pp. 1-13 ◽  
Author(s):  
Suryo Adhi Wibowo ◽  
Hansoo Lee ◽  
Eun Kyeong Kim ◽  
Sungshin Kim

The representation of the object is an important factor in building a robust visual object tracking algorithm. To resolve this problem, complementary learners that use color histogram- and correlation filter-based representation to represent the target object can be used since they each have advantages that can be exploited to compensate the other’s drawback in visual tracking. Further, a tracking algorithm can fail because of the distractor, even when complementary learners have been implemented for the target object representation. In this study, we show that, in order to handle the distractor, first the distractor must be detected by learning the responses from the color-histogram- and correlation-filter-based representation. Then, to determine the target location, we can decide whether the responses from each representation should be merged or only the response from the correlation filter should be used. This decision depends on the result obtained from the distractor detection process. Experiments were performed on the widely used VOT2014 and VOT2015 benchmark datasets. It was verified that our proposed method performs favorably as compared with several state-of-the-art visual tracking algorithms.


Electronics ◽  
2020 ◽  
Vol 9 (11) ◽  
pp. 1918 ◽  
Author(s):  
Md. Maklachur Rahman ◽  
Md Rishad Ahmed ◽  
Lamyanba Laishram ◽  
Seock Ho Kim ◽  
Soon Ki Jung

Siamese network-based trackers are broadly applied to solve visual tracking problems due to its balanced performance in terms of speed and accuracy. Tracking desired objects in challenging scenarios is still one of the fundamental concerns during visual tracking. This research paper proposes a feature refined end-to-end tracking framework with real-time tracking speed and considerable performance. The feature refine network has been incorporated to enhance the target feature representation power, utilizing high-level semantic information. Besides, it allows the network to capture the salient information to locate the target and learns to represent the target feature in a more generalized way advancing the overall tracking performance, particularly in the challenging sequences. But, only the feature refine module is unable to handle such challenges because of its less discriminative ability. To overcome this difficulty, we employ an attention module inside the feature refine network that strengths the tracker discrimination ability between the target and background. Furthermore, we conduct extensive experiments to ensure the proposed tracker’s effectiveness using several popular tracking benchmarks, demonstrating that our proposed model achieves state-of-the-art performance over other trackers.


Author(s):  
Rawan Fayez ◽  
Mohamed Taha Abd Elfattah Taha ◽  
Mahmoud Gadallah

Visual object tracking remains a challenge facing an intelligent control system. A variety of applications serve many purposes such as surveillance. The developed technology faces plenty of obstacles that should be addressed including occlusion. In visual tracking, online learning techniques are most common due to their efficiency for most video sequences. Many object tracking techniques have emerged. However, the drifting problem in the case of noisy updates has been a stumbling block for the majority of relevant techniques. Such a problem can now be surmounted through updating the classifiers. The proposed system is called the Occluded Object Tracking System (OOTS) It is a hybrid system constructed from two algorithms: a fast technique Circulant Structure Kernels with Color Names (CSK-CN) and an efficient algorithm occlusion-aware Real-time Object Tracking (ROT). The proposed OOTS is evaluated with standard visual tracking benchmark databases. The experimental results proved that the proposed OOTS system is more reliable and provides efficient tracking results than other compared methods.


Sensors ◽  
2019 ◽  
Vol 19 (2) ◽  
pp. 387 ◽  
Author(s):  
Ming Du ◽  
Yan Ding ◽  
Xiuyun Meng ◽  
Hua-Liang Wei ◽  
Yifan Zhao

In recent years, regression trackers have drawn increasing attention in the visual-object tracking community due to their favorable performance and easy implementation. The tracker algorithms directly learn mapping from dense samples around the target object to Gaussian-like soft labels. However, in many real applications, when applied to test data, the extreme imbalanced distribution of training samples usually hinders the robustness and accuracy of regression trackers. In this paper, we propose a novel effective distractor-aware loss function to balance this issue by highlighting the significant domain and by severely penalizing the pure background. In addition, we introduce a full differentiable hierarchy-normalized concatenation connection to exploit abstractions across multiple convolutional layers. Extensive experiments were conducted on five challenging benchmark-tracking datasets, that is, OTB-13, OTB-15, TC-128, UAV-123, and VOT17. The experimental results are promising and show that the proposed tracker performs much better than nearly all the compared state-of-the-art approaches.


Algorithms ◽  
2018 ◽  
Vol 12 (1) ◽  
pp. 8 ◽  
Author(s):  
Wancheng Zhang ◽  
Yanmin Luo ◽  
Zhi Chen ◽  
Yongzhao Du ◽  
Daxin Zhu ◽  
...  

Discriminative correlation filters (DCFs) have been shown to perform superiorly in visual object tracking. However, visual tracking is still challenging when the target objects undergo complex scenarios such as occlusion, deformation, scale changes and illumination changes. In this paper, we utilize the hierarchical features of convolutional neural networks (CNNs) and learn a spatial-temporal context correlation filter on convolutional layers. Then, the translation is estimated by fusing the response score of the filters on the three convolutional layers. In terms of scale estimation, we learn a discriminative correlation filter to estimate scale from the best confidence results. Furthermore, we proposed a re-detection activation discrimination method to improve the robustness of visual tracking in the case of tracking failure and an adaptive model update method to reduce tracking drift caused by noisy updates. We evaluate the proposed tracker with DCFs and deep features on OTB benchmark datasets. The tracking results demonstrated that the proposed algorithm is superior to several state-of-the-art DCF methods in terms of accuracy and robustness.


Sign in / Sign up

Export Citation Format

Share Document