Quality Matters: Embracing Quality Clues for Robust 3D Multi-Object
Tracking
- URL: http://arxiv.org/abs/2208.10976v1
- Date: Tue, 23 Aug 2022 13:47:14 GMT
- Title: Quality Matters: Embracing Quality Clues for Robust 3D Multi-Object
Tracking
- Authors: Jinrong Yang, En Yu, Zeming Li, Xiaoping Li, Wenbing Tao
- Abstract summary: 3D Multi-Object Tracking (MOT) has achieved tremendous achievement thanks to the rapid development of 3D object detection and 2D MOT.
Recent advanced works generally employ a series of object attributes, e.g., position, size, velocity, and appearance, to provide the clues for the association in 3D MOT.
We propose a quality-aware object association strategy to leverage the quality score as an important reference factor for achieving robust association.
- Score: 30.47607974875308
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D Multi-Object Tracking (MOT) has achieved tremendous achievement thanks to
the rapid development of 3D object detection and 2D MOT. Recent advanced works
generally employ a series of object attributes, e.g., position, size, velocity,
and appearance, to provide the clues for the association in 3D MOT. However,
these cues may not be reliable due to some visual noise, such as occlusion and
blur, leading to tracking performance bottleneck. To reveal the dilemma, we
conduct extensive empirical analysis to expose the key bottleneck of each clue
and how they correlate with each other. The analysis results motivate us to
efficiently absorb the merits among all cues, and adaptively produce an optimal
tacking manner. Specifically, we present Location and Velocity Quality
Learning, which efficiently guides the network to estimate the quality of
predicted object attributes. Based on these quality estimations, we propose a
quality-aware object association (QOA) strategy to leverage the quality score
as an important reference factor for achieving robust association. Despite its
simplicity, extensive experiments indicate that the proposed strategy
significantly boosts tracking performance by 2.2% AMOTA and our method
outperforms all existing state-of-the-art works on nuScenes by a large margin.
Moreover, QTrack achieves 48.0% and 51.1% AMOTA tracking performance on the
nuScenes validation and test sets, which significantly reduces the performance
gap between pure camera and LiDAR based trackers.
Related papers
- Temporal Correlation Meets Embedding: Towards a 2nd Generation of JDE-based Real-Time Multi-Object Tracking [52.04679257903805]
Joint Detection and Embedding(JDE) trackers have demonstrated excellent performance in Multi-Object Tracking(MOT) tasks.
We propose a new learning approach using cross-correlation to capture temporal information of objects.
Our tracker, named TCBTrack, achieves state-of-the-art performance on multiple public benchmarks.
arXiv Detail & Related papers (2024-07-19T07:48:45Z) - KECOR: Kernel Coding Rate Maximization for Active 3D Object Detection [48.66703222700795]
We resort to a novel kernel strategy to identify the most informative point clouds to acquire labels.
To accommodate both one-stage (i.e., SECOND) and two-stage detectors, we incorporate the classification entropy tangent and well trade-off between detection performance and the total number of bounding boxes selected for annotation.
Our results show that approximately 44% box-level annotation costs and 26% computational time are reduced compared to the state-of-the-art method.
arXiv Detail & Related papers (2023-07-16T04:27:03Z) - Joint Feature Learning and Relation Modeling for Tracking: A One-Stream
Framework [76.70603443624012]
We propose a novel one-stream tracking (OSTrack) framework that unifies feature learning and relation modeling.
In this way, discriminative target-oriented features can be dynamically extracted by mutual guidance.
OSTrack achieves state-of-the-art performance on multiple benchmarks, in particular, it shows impressive results on the one-shot tracking benchmark GOT-10k.
arXiv Detail & Related papers (2022-03-22T18:37:11Z) - Distractor-Aware Fast Tracking via Dynamic Convolutions and MOT
Philosophy [63.91005999481061]
A practical long-term tracker typically contains three key properties, i.e. an efficient model design, an effective global re-detection strategy and a robust distractor awareness mechanism.
We propose a two-task tracking frame work (named DMTrack) to achieve distractor-aware fast tracking via Dynamic convolutions (d-convs) and Multiple object tracking (MOT) philosophy.
Our tracker achieves state-of-the-art performance on the LaSOT, OxUvA, TLP, VOT2018LT and VOT 2019LT benchmarks and runs in real-time (3x faster
arXiv Detail & Related papers (2021-04-25T00:59:53Z) - Learnable Online Graph Representations for 3D Multi-Object Tracking [156.58876381318402]
We propose a unified and learning based approach to the 3D MOT problem.
We employ a Neural Message Passing network for data association that is fully trainable.
We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.
arXiv Detail & Related papers (2021-04-23T17:59:28Z) - Robust Ego and Object 6-DoF Motion Estimation and Tracking [5.162070820801102]
This paper proposes a robust solution to achieve accurate estimation and consistent track-ability for dynamic multi-body visual odometry.
A compact and effective framework is proposed leveraging recent advances in semantic instance-level segmentation and accurate optical flow estimation.
A novel formulation, jointly optimizing SE(3) motion and optical flow is introduced that improves the quality of the tracked points and the motion estimation accuracy.
arXiv Detail & Related papers (2020-07-28T05:12:56Z) - SESS: Self-Ensembling Semi-Supervised 3D Object Detection [138.80825169240302]
We propose SESS, a self-ensembling semi-supervised 3D object detection framework. Specifically, we design a thorough perturbation scheme to enhance generalization of the network on unlabeled and new unseen data.
Our SESS achieves competitive performance compared to the state-of-the-art fully-supervised method by using only 50% labeled data.
arXiv Detail & Related papers (2019-12-26T08:48:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.