The Multi-task Fully Convolutional Siamese Network with Correlation Filter Layer for Real-Time Visual Tracking

Author(s):  
Shiyu Xuan ◽  
Shengyang Li ◽  
Zifei Zhao ◽  
Mingfei Han
2019 ◽  
Vol 55 (13) ◽  
pp. 742-745 ◽  
Author(s):  
Kang Yang ◽  
Huihui Song ◽  
Kaihua Zhang ◽  
Jiaqing Fan

Sensors ◽  
2019 ◽  
Vol 19 (22) ◽  
pp. 4904 ◽  
Author(s):  
Yeongbin Kim ◽  
Joongchol Shin ◽  
Hasil Park ◽  
Joonki Paik

Online training framework based on discriminative correlation filters for visual tracking has recently shown significant improvement in both accuracy and speed. However, correlation filter-base discriminative approaches have a common problem of tracking performance degradation when the local structure of a target is distorted by the boundary effect problem. The shape distortion of the target is mainly caused by the circulant structure in the Fourier domain processing, and it makes the correlation filter learn distorted training samples. In this paper, we present a structure–attention network to preserve the target structure from the structure distortion caused by the boundary effect. More specifically, we adopt a variational auto-encoder as a structure–attention network to make various and representative target structures. We also proposed two denoising criteria using a novel reconstruction loss for variational auto-encoding framework to capture more robust structures even under the boundary condition. Through the proposed structure–attention framework, discriminative correlation filters can learn robust structure information of targets during online training with an enhanced discriminating performance and adaptability. Experimental results on major visual tracking benchmark datasets show that the proposed method produces a better or comparable performance compared with the state-of-the-art tracking methods with a real-time processing speed of more than 80 frames per second.


Author(s):  
Yunhua Zhang ◽  
Lijun Wang ◽  
Jinqing Qi ◽  
Dong Wang ◽  
Mengyang Feng ◽  
...  

2020 ◽  
Vol 103 ◽  
pp. 104002 ◽  
Author(s):  
Mu Zhu ◽  
Hui Zhang ◽  
Jing Zhang ◽  
Li Zhuo

Author(s):  
Lei Pu ◽  
Xinxi Feng ◽  
Zhiqiang Hou ◽  
Wangsheng Yu ◽  
Yufei Zha

Author(s):  
Xiaoliang Wang ◽  
Marie O'Brien ◽  
Changle Xiang ◽  
Bin Xu ◽  
Homayoun Najjaran

Author(s):  
Lei Pu ◽  
Xinxi Feng ◽  
Zhiqiang Hou ◽  
Wangsheng Yu ◽  
Yufei Zha ◽  
...  

2019 ◽  
Vol 32 (18) ◽  
pp. 14335-14346 ◽  
Author(s):  
Kang Yang ◽  
Huihui Song ◽  
Kaihua Zhang ◽  
Qingshan Liu

2020 ◽  
Vol 2020 ◽  
pp. 1-19
Author(s):  
Chenpu Li ◽  
Qianjian Xing ◽  
Zhenguo Ma ◽  
Ke Zang

With the development of deep learning, trackers based on convolutional neural networks (CNNs) have made significant achievements in visual tracking over the years. The fully connected Siamese network (SiamFC) is a typical representation of those trackers. SiamFC designs a two-branch architecture of a CNN and models’ visual tracking as a general similarity-learning problem. However, the feature maps it uses for visual tracking are only from the last layer of the CNN. Those features contain high-level semantic information but lack sufficiently detailed texture information. This means that the SiamFC tracker tends to drift when there are other same-category objects or when the contrast between the target and the background is very low. Focusing on addressing this problem, we design a novel tracking algorithm that combines a correlation filter tracker and the SiamFC tracker into one framework. In this framework, the correlation filter tracker can use the Histograms of Oriented Gradients (HOG) and color name (CN) features to guide the SiamFC tracker. This framework also contains an evaluation criterion which we design to evaluate the tracking result of the two trackers. If this criterion finds the SiamFC tracker fails in some cases, our framework will use the tracking result from the correlation filter tracker to correct the SiamFC. In this way, the defects of SiamFC’s high-level semantic features are remedied by the HOG and CN features. So, our algorithm provides a framework which combines two trackers together and makes them complement each other in visual tracking. And to the best of our knowledge, our algorithm is also the first one which designs an evaluation criterion using correlation filter and zero padding to evaluate the tracking result. Comprehensive experiments are conducted on the Online Tracking Benchmark (OTB), Temple Color (TC128), Benchmark for UAV Tracking (UAV-123), and Visual Object Tracking (VOT) Benchmark. The results show that our algorithm achieves quite a competitive performance when compared with the baseline tracker and several other state-of-the-art trackers.


2019 ◽  
Vol 127 ◽  
pp. 138-145 ◽  
Author(s):  
Lei Qu ◽  
Kuixiang Liu ◽  
Baochen Yao ◽  
Jun Tang ◽  
Wei Zhang

Sign in / Sign up

Export Citation Format

Share Document