TLPG-Tracker: Joint Learning of Target Localization and Proposal Generation for Visual Tracking

TLPG-Tracker: Joint Learning of Target Localization and Proposal Generation for Visual Tracking

Siyuan Li, Zhi Zhang, Ziyu Liu, Anna Wang, Linglong Qiu, Feng Du

Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence
Main track. Pages 708-715. https://doi.org/10.24963/ijcai.2020/99

Target localization and proposal generation are two essential subtasks in generic visual tracking, and it is a challenge to address both the two efficiently. In this paper, we propose an efficient two-stage architecture which makes full use of the complementarity of two subtasks to achieve robust localization and high-quality proposals generation of the target jointly. Specifically, our model performs a novel deformable central correlation operation by an online learning model in both two stages to locate new target centers while generating target proposals in the vicinity of these centers. The proposals are refined in the refinement stage to further improve accuracy and robustness. Moreover, the model benefits from multi-level features aggregation in a neck module and a feature enhancement module. We conduct extensive ablation studies to demonstrate the effectiveness of our proposed methods. Our tracker runs at over 30 FPS and sets a new state-of-the-art on five tracking benchmarks, including LaSOT, VOT2018, TrackingNet, GOT10k, OTB2015.
Keywords:
Computer Vision: Motion and Tracking
Machine Learning: Deep Learning: Convolutional networks
Computer Vision: Video: Events, Activities and Surveillance
Machine Learning: Deep Learning