Online Multiple Object Tracking with Cross-Task Synergy
- URL: http://arxiv.org/abs/2104.00380v1
- Date: Thu, 1 Apr 2021 10:19:40 GMT
- Title: Online Multiple Object Tracking with Cross-Task Synergy
- Authors: Song Guo, Jingya Wang, Xinchao Wang, Dacheng Tao
- Abstract summary: We propose a novel unified model with synergy between position prediction and embedding association.
The two tasks are linked by temporal-aware target attention and distractor attention, as well as identity-aware memory aggregation model.
- Score: 120.70085565030628
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern online multiple object tracking (MOT) methods usually focus on two
directions to improve tracking performance. One is to predict new positions in
an incoming frame based on tracking information from previous frames, and the
other is to enhance data association by generating more discriminative identity
embeddings. Some works combined both directions within one framework but
handled them as two individual tasks, thus gaining little mutual benefits. In
this paper, we propose a novel unified model with synergy between position
prediction and embedding association. The two tasks are linked by
temporal-aware target attention and distractor attention, as well as
identity-aware memory aggregation model. Specifically, the attention modules
can make the prediction focus more on targets and less on distractors,
therefore more reliable embeddings can be extracted accordingly for
association. On the other hand, such reliable embeddings can boost
identity-awareness through memory aggregation, hence strengthen attention
modules and suppress drifts. In this way, the synergy between position
prediction and embedding association is achieved, which leads to strong
robustness to occlusions. Extensive experiments demonstrate the superiority of
our proposed model over a wide range of existing methods on MOTChallenge
benchmarks. Our code and models are publicly available at
https://github.com/songguocode/TADAM.
Related papers
- STCMOT: Spatio-Temporal Cohesion Learning for UAV-Based Multiple Object Tracking [13.269416985959404]
Multiple object tracking (MOT) in Unmanned Aerial Vehicle (UAV) videos is important for diverse applications in computer vision.
We propose a novel Spatio-Temporal Cohesion Multiple Object Tracking framework (STCMOT)
We use historical embedding features to model the representation of ReID and detection features in a sequential order.
Our framework sets a new state-of-the-art performance in MOTA and IDF1 metrics.
arXiv Detail & Related papers (2024-09-17T14:34:18Z) - Single-Shot and Multi-Shot Feature Learning for Multi-Object Tracking [55.13878429987136]
We propose a simple yet effective two-stage feature learning paradigm to jointly learn single-shot and multi-shot features for different targets.
Our method has achieved significant improvements on MOT17 and MOT20 datasets while reaching state-of-the-art performance on DanceTrack dataset.
arXiv Detail & Related papers (2023-11-17T08:17:49Z) - Modeling Continuous Motion for 3D Point Cloud Object Tracking [54.48716096286417]
This paper presents a novel approach that views each tracklet as a continuous stream.
At each timestamp, only the current frame is fed into the network to interact with multi-frame historical features stored in a memory bank.
To enhance the utilization of multi-frame features for robust tracking, a contrastive sequence enhancement strategy is proposed.
arXiv Detail & Related papers (2023-03-14T02:58:27Z) - 3DMODT: Attention-Guided Affinities for Joint Detection & Tracking in 3D
Point Clouds [95.54285993019843]
We propose a method for joint detection and tracking of multiple objects in 3D point clouds.
Our model exploits temporal information employing multiple frames to detect objects and track them in a single network.
arXiv Detail & Related papers (2022-11-01T20:59:38Z) - Joint Feature Learning and Relation Modeling for Tracking: A One-Stream
Framework [76.70603443624012]
We propose a novel one-stream tracking (OSTrack) framework that unifies feature learning and relation modeling.
In this way, discriminative target-oriented features can be dynamically extracted by mutual guidance.
OSTrack achieves state-of-the-art performance on multiple benchmarks, in particular, it shows impressive results on the one-shot tracking benchmark GOT-10k.
arXiv Detail & Related papers (2022-03-22T18:37:11Z) - Correlation-Aware Deep Tracking [83.51092789908677]
We propose a novel target-dependent feature network inspired by the self-/cross-attention scheme.
Our network deeply embeds cross-image feature correlation in multiple layers of the feature network.
Our model can be flexibly pre-trained on abundant unpaired images, leading to notably faster convergence than the existing methods.
arXiv Detail & Related papers (2022-03-03T11:53:54Z) - Rethinking the competition between detection and ReID in Multi-Object
Tracking [44.59367033562385]
One-shot models which jointly learn detection and identification embeddings, have drawn great attention in multi-object tracking (MOT)
In this paper, we propose a novel reciprocal network (REN) with a self-relation and cross-relation design to better learn task-dependent representations.
We also introduce a scale-aware attention network (SAAN) that prevents semantic level misalignment to improve the association capability of ID embeddings.
arXiv Detail & Related papers (2020-10-23T02:44:59Z) - Joint Object Detection and Multi-Object Tracking with Graph Neural
Networks [32.1359455541169]
We propose a new instance of joint MOT approach based on Graph Neural Networks (GNNs)
We show the effectiveness of our GNN-based joint MOT approach and show state-of-the-art performance for both detection and MOT tasks.
arXiv Detail & Related papers (2020-06-23T17:07:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.