Target-Aware Tracking with Long-term Context Attention
- URL: http://arxiv.org/abs/2302.13840v1
- Date: Mon, 27 Feb 2023 14:40:58 GMT
- Title: Target-Aware Tracking with Long-term Context Attention
- Authors: Kaijie He, Canlong Zhang, Sheng Xie, Zhixin Li, Zhiwen Wang
- Abstract summary: Long-term context attention (LCA) module can perform extensive information fusion on the target and its context from long-term frames.
LCA uses the target state from the previous frame to exclude the interference of similar objects and complex backgrounds.
Our tracker achieves state-of-the-art performance on multiple benchmarks, with 71.1% AUC, 89.3% NP, and 73.0% AO on LaSOT, TrackingNet, and GOT-10k.
- Score: 8.20858704675519
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most deep trackers still follow the guidance of the siamese paradigms and use
a template that contains only the target without any contextual information,
which makes it difficult for the tracker to cope with large appearance changes,
rapid target movement, and attraction from similar objects. To alleviate the
above problem, we propose a long-term context attention (LCA) module that can
perform extensive information fusion on the target and its context from
long-term frames, and calculate the target correlation while enhancing target
features. The complete contextual information contains the location of the
target as well as the state around the target. LCA uses the target state from
the previous frame to exclude the interference of similar objects and complex
backgrounds, thus accurately locating the target and enabling the tracker to
obtain higher robustness and regression accuracy. By embedding the LCA module
in Transformer, we build a powerful online tracker with a target-aware
backbone, termed as TATrack. In addition, we propose a dynamic online update
algorithm based on the classification confidence of historical information
without additional calculation burden. Our tracker achieves state-of-the-art
performance on multiple benchmarks, with 71.1\% AUC, 89.3\% NP, and 73.0\% AO
on LaSOT, TrackingNet, and GOT-10k. The code and trained models are available
on https://github.com/hekaijie123/TATrack.
Related papers
- RTracker: Recoverable Tracking via PN Tree Structured Memory [71.05904715104411]
We propose a recoverable tracking framework, RTracker, that uses a tree-structured memory to dynamically associate a tracker and a detector to enable self-recovery.
Specifically, we propose a Positive-Negative Tree-structured memory to chronologically store and maintain positive and negative target samples.
Our core idea is to use the support samples of positive and negative target categories to establish a relative distance-based criterion for a reliable assessment of target loss.
arXiv Detail & Related papers (2024-03-28T08:54:40Z) - Autoregressive Queries for Adaptive Tracking with Spatio-TemporalTransformers [55.46413719810273]
rich-temporal information is crucial to the complicated target appearance in visual tracking.
Our method improves the tracker's performance on six popular tracking benchmarks.
arXiv Detail & Related papers (2024-03-15T02:39:26Z) - Context-aware Visual Tracking with Joint Meta-updating [11.226947525556813]
We propose a context-aware tracking model to optimize the tracker over the representation space, which jointly meta-update both branches by exploiting information along the whole sequence.
The proposed tracking method achieves an EAO score of 0.514 on VOT2018 with the speed of 40FPS, demonstrating its capability of improving the accuracy and robustness of the underlying tracker with little speed drop.
arXiv Detail & Related papers (2022-04-04T14:16:00Z) - Learning Dynamic Compact Memory Embedding for Deformable Visual Object
Tracking [82.34356879078955]
We propose a compact memory embedding to enhance the discrimination of the segmentation-based deformable visual tracking method.
Our method outperforms the excellent segmentation-based trackers, i.e., D3S and SiamMask on DAVIS 2017 benchmark.
arXiv Detail & Related papers (2021-11-23T03:07:12Z) - Target Transformed Regression for Accurate Tracking [30.516462193231888]
This paper repurposes a Transformer-alike regression branch, termed as Target Transformed Regression (TREG) for accurate anchor-free tracking.
The core to our TREG is to model pair-wise relation between elements in target template and search region, and use the resulted target enhanced visual representation for accurate bounding box regression.
In addition, we devise a simple online template update mechanism to select reliable templates, increasing the robustness for appearance variations and geometric deformations of target in time.
arXiv Detail & Related papers (2021-04-01T11:25:23Z) - STMTrack: Template-free Visual Tracking with Space-time Memory Networks [42.06375415765325]
Existing trackers with template updating mechanisms rely on time-consuming numerical optimization and complex hand-designed strategies to achieve competitive performance.
We propose a novel tracking framework built on top of a space-time memory network that is competent to make full use of historical information related to the target.
Specifically, a novel memory mechanism is introduced, which stores the historical information of the target to guide the tracker to focus on the most informative regions in the current frame.
arXiv Detail & Related papers (2021-04-01T08:10:56Z) - Learning Target Candidate Association to Keep Track of What Not to Track [100.80610986625693]
We propose to keep track of distractor objects in order to continue tracking the target.
To tackle the problem of lacking ground-truth correspondences between distractor objects in visual tracking, we propose a training strategy that combines partial annotations with self-supervision.
Our tracker sets a new state-of-the-art on six benchmarks, achieving an AUC score of 67.2% on LaSOT and a +6.1% absolute gain on the OxUvA long-term dataset.
arXiv Detail & Related papers (2021-03-30T17:58:02Z) - TDIOT: Target-driven Inference for Deep Video Object Tracking [0.2457872341625575]
In this work, we adopt the pre-trained Mask R-CNN deep object detector as the baseline.
We introduce a novel inference architecture placed on top of FPN-ResNet101 backbone of Mask R-CNN to jointly perform detection and tracking.
The proposed single object tracker, TDIOT, applies an appearance similarity-based temporal matching for data association.
arXiv Detail & Related papers (2021-03-19T20:45:06Z) - Graph Attention Tracking [76.19829750144564]
We propose a simple target-aware Siamese graph attention network for general object tracking.
Experiments on challenging benchmarks including GOT-10k, UAV123, OTB-100 and LaSOT demonstrate that the proposed SiamGAT outperforms many state-of-the-art trackers.
arXiv Detail & Related papers (2020-11-23T04:26:45Z) - Ocean: Object-aware Anchor-free Tracking [75.29960101993379]
The regression network in anchor-based methods is only trained on the positive anchor boxes.
We propose a novel object-aware anchor-free network to address this issue.
Our anchor-free tracker achieves state-of-the-art performance on five benchmarks.
arXiv Detail & Related papers (2020-06-18T17:51:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.