Target-Aware Tracking with Long-Term Context Attention

Authors

  • Kaijie He School of Computer Science and Engineering, Guangxi Normal University, China
  • Canlong Zhang School of Computer Science and Engineering, Guangxi Normal University, China Guangxi Key Lab of Multi-source Information Mining and Security, China
  • Sheng Xie School of Computer Science and Engineering, Guangxi Normal University, China
  • Zhixin Li School of Computer Science and Engineering, Guangxi Normal University, China Guangxi Key Lab of Multi-source Information Mining and Security, China
  • Zhiwen Wang School of Computer Science and Technology, Guangxi University of Science and Technology, China

DOI:

https://doi.org/10.1609/aaai.v37i1.25155

Keywords:

CV: Motion & Tracking, CV: Applications, CV: Other Foundations of Computer Vision

Abstract

Most deep trackers still follow the guidance of the siamese paradigms and use a template that contains only the target without any contextual information, which makes it difficult for the tracker to cope with large appearance changes, rapid target movement, and attraction from similar objects. To alleviate the above problem, we propose a long-term context attention (LCA) module that can perform extensive information fusion on the target and its context from long-term frames, and calculate the target correlation while enhancing target features. The complete contextual information contains the location of the target as well as the state around the target. LCA uses the target state from the previous frame to exclude the interference of similar objects and complex backgrounds, thus accurately locating the target and enabling the tracker to obtain higher robustness and regression accuracy. By embedding the LCA module in Transformer, we build a powerful online tracker with a target-aware backbone, termed as TATrack. In addition, we propose a dynamic online update algorithm based on the classification confidence of historical information without additional calculation burden. Our tracker achieves state-of-the-art performance on multiple benchmarks, with 71.1% AUC, 89.3% NP, and 73.0% AO on LaSOT, TrackingNet, and GOT-10k. The code and trained models are available on https://github.com/hekaijie123/TATrack.

Downloads

Published

2023-06-26

How to Cite

He, K., Zhang, C., Xie, S., Li, Z., & Wang, Z. (2023). Target-Aware Tracking with Long-Term Context Attention. Proceedings of the AAAI Conference on Artificial Intelligence, 37(1), 773-780. https://doi.org/10.1609/aaai.v37i1.25155

Issue

Section

AAAI Technical Track on Computer Vision I