Learning Global Structure Consistency for Robust Object Tracking
- URL: http://arxiv.org/abs/2008.11769v1
- Date: Wed, 26 Aug 2020 19:12:53 GMT
- Title: Learning Global Structure Consistency for Robust Object Tracking
- Authors: Bi Li, Chengquan Zhang, Zhibin Hong, Xu Tang, Jingtuo Liu, Junyu Han,
Errui Ding, Wenyu Liu
- Abstract summary: This work considers the emphtransient variations of the whole scene.
We propose an effective and efficient short-term model that learns to exploit the global structure consistency in a short time.
We empirically verify that the proposed tracker can tackle the two challenging scenarios and validate it on large scale benchmarks.
- Score: 57.736915865309165
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fast appearance variations and the distractions of similar objects are two of
the most challenging problems in visual object tracking. Unlike many existing
trackers that focus on modeling only the target, in this work, we consider the
\emph{transient variations of the whole scene}. The key insight is that the
object correspondence and spatial layout of the whole scene are consistent
(i.e., global structure consistency) in consecutive frames which helps to
disambiguate the target from distractors. Moreover, modeling transient
variations enables to localize the target under fast variations. Specifically,
we propose an effective and efficient short-term model that learns to exploit
the global structure consistency in a short time and thus can handle fast
variations and distractors. Since short-term modeling falls short of handling
occlusion and out of the views, we adopt the long-short term paradigm and use a
long-term model that corrects the short-term model when it drifts away from the
target or the target is not present. These two components are carefully
combined to achieve the balance of stability and plasticity during tracking. We
empirically verify that the proposed tracker can tackle the two challenging
scenarios and validate it on large scale benchmarks. Remarkably, our tracker
improves state-of-the-art-performance on VOT2018 from 0.440 to 0.460, GOT-10k
from 0.611 to 0.640, and NFS from 0.619 to 0.629.
Related papers
- Exploring Dynamic Transformer for Efficient Object Tracking [58.120191254379854]
We propose DyTrack, a dynamic transformer framework for efficient tracking.
DyTrack automatically learns to configure proper reasoning routes for various inputs, gaining better utilization of the available computational budget.
Experiments on multiple benchmarks demonstrate that DyTrack achieves promising speed-precision trade-offs with only a single model.
arXiv Detail & Related papers (2024-03-26T12:31:58Z) - MotionTrack: Learning Robust Short-term and Long-term Motions for
Multi-Object Tracking [56.92165669843006]
We propose MotionTrack, which learns robust short-term and long-term motions in a unified framework to associate trajectories from a short to long range.
For dense crowds, we design a novel Interaction Module to learn interaction-aware motions from short-term trajectories, which can estimate the complex movement of each target.
For extreme occlusions, we build a novel Refind Module to learn reliable long-term motions from the target's history trajectory, which can link the interrupted trajectory with its corresponding detection.
arXiv Detail & Related papers (2023-03-18T12:38:33Z) - Target-Aware Tracking with Long-term Context Attention [8.20858704675519]
Long-term context attention (LCA) module can perform extensive information fusion on the target and its context from long-term frames.
LCA uses the target state from the previous frame to exclude the interference of similar objects and complex backgrounds.
Our tracker achieves state-of-the-art performance on multiple benchmarks, with 71.1% AUC, 89.3% NP, and 73.0% AO on LaSOT, TrackingNet, and GOT-10k.
arXiv Detail & Related papers (2023-02-27T14:40:58Z) - Context-aware Visual Tracking with Joint Meta-updating [11.226947525556813]
We propose a context-aware tracking model to optimize the tracker over the representation space, which jointly meta-update both branches by exploiting information along the whole sequence.
The proposed tracking method achieves an EAO score of 0.514 on VOT2018 with the speed of 40FPS, demonstrating its capability of improving the accuracy and robustness of the underlying tracker with little speed drop.
arXiv Detail & Related papers (2022-04-04T14:16:00Z) - Distractor-Aware Fast Tracking via Dynamic Convolutions and MOT
Philosophy [63.91005999481061]
A practical long-term tracker typically contains three key properties, i.e. an efficient model design, an effective global re-detection strategy and a robust distractor awareness mechanism.
We propose a two-task tracking frame work (named DMTrack) to achieve distractor-aware fast tracking via Dynamic convolutions (d-convs) and Multiple object tracking (MOT) philosophy.
Our tracker achieves state-of-the-art performance on the LaSOT, OxUvA, TLP, VOT2018LT and VOT 2019LT benchmarks and runs in real-time (3x faster
arXiv Detail & Related papers (2021-04-25T00:59:53Z) - Target Transformed Regression for Accurate Tracking [30.516462193231888]
This paper repurposes a Transformer-alike regression branch, termed as Target Transformed Regression (TREG) for accurate anchor-free tracking.
The core to our TREG is to model pair-wise relation between elements in target template and search region, and use the resulted target enhanced visual representation for accurate bounding box regression.
In addition, we devise a simple online template update mechanism to select reliable templates, increasing the robustness for appearance variations and geometric deformations of target in time.
arXiv Detail & Related papers (2021-04-01T11:25:23Z) - Learning Target Candidate Association to Keep Track of What Not to Track [100.80610986625693]
We propose to keep track of distractor objects in order to continue tracking the target.
To tackle the problem of lacking ground-truth correspondences between distractor objects in visual tracking, we propose a training strategy that combines partial annotations with self-supervision.
Our tracker sets a new state-of-the-art on six benchmarks, achieving an AUC score of 67.2% on LaSOT and a +6.1% absolute gain on the OxUvA long-term dataset.
arXiv Detail & Related papers (2021-03-30T17:58:02Z) - Learning to Track with Object Permanence [61.36492084090744]
We introduce an end-to-end trainable approach for joint object detection and tracking.
Our model, trained jointly on synthetic and real data, outperforms the state of the art on KITTI, and MOT17 datasets.
arXiv Detail & Related papers (2021-03-26T04:43:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.