An Effective Motion-Centric Paradigm for 3D Single Object Tracking in
Point Clouds
- URL: http://arxiv.org/abs/2303.12535v2
- Date: Thu, 12 Oct 2023 02:38:50 GMT
- Title: An Effective Motion-Centric Paradigm for 3D Single Object Tracking in
Point Clouds
- Authors: Chaoda Zheng, Xu Yan, Haiming Zhang, Baoyuan Wang, Shenghui Cheng,
Shuguang Cui, Zhen Li
- Abstract summary: 3D single object tracking in LiDAR point clouds (LiDAR SOT) plays a crucial role in autonomous driving.
Current approaches all follow the Siamese paradigm based on appearance matching.
We introduce a motion-centric paradigm to handle LiDAR SOT from a new perspective.
- Score: 50.19288542498838
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D single object tracking in LiDAR point clouds (LiDAR SOT) plays a crucial
role in autonomous driving. Current approaches all follow the Siamese paradigm
based on appearance matching. However, LiDAR point clouds are usually
textureless and incomplete, which hinders effective appearance matching.
Besides, previous methods greatly overlook the critical motion clues among
targets. In this work, beyond 3D Siamese tracking, we introduce a
motion-centric paradigm to handle LiDAR SOT from a new perspective. Following
this paradigm, we propose a matching-free two-stage tracker M^2-Track. At the
1st-stage, M^2-Track localizes the target within successive frames via motion
transformation. Then it refines the target box through motion-assisted shape
completion at the 2nd-stage. Due to the motion-centric nature, our method shows
its impressive generalizability with limited training labels and provides good
differentiability for end-to-end cycle training. This inspires us to explore
semi-supervised LiDAR SOT by incorporating a pseudo-label-based motion
augmentation and a self-supervised loss term. Under the fully-supervised
setting, extensive experiments confirm that M^2-Track significantly outperforms
previous state-of-the-arts on three large-scale datasets while running at 57FPS
(~3%, ~11% and ~22% precision gains on KITTI, NuScenes, and Waymo Open Dataset
respectively). While under the semi-supervised setting, our method performs on
par with or even surpasses its fully-supervised counterpart using fewer than
half of the labels from KITTI. Further analysis verifies each component's
effectiveness and shows the motion-centric paradigm's promising potential for
auto-labeling and unsupervised domain adaptation.
Related papers
- BiTrack: Bidirectional Offline 3D Multi-Object Tracking Using Camera-LiDAR Data [11.17376076195671]
"BiTrack" is a 3D OMOT framework that includes modules of 2D-3D detection fusion, initial trajectory generation, and bidirectional trajectory re-optimization.
The experiment results on the KITTI dataset demonstrate that BiTrack achieves the state-of-the-art performance for 3D OMOT tasks in terms of accuracy and efficiency.
arXiv Detail & Related papers (2024-06-26T15:09:54Z) - Motion-to-Matching: A Mixed Paradigm for 3D Single Object Tracking [27.805298263103495]
We propose MTM-Tracker, which combines motion modeling with feature matching into a single network.
In the first stage, we exploit the continuous historical boxes as motion prior and propose an encoder-decoder structure to locate target coarsely.
In the second stage, we introduce a feature interaction module to extract motion-aware features from consecutive point clouds and match them to refine target movement.
arXiv Detail & Related papers (2023-08-23T02:40:51Z) - MotionTrack: Learning Motion Predictor for Multiple Object Tracking [68.68339102749358]
We introduce a novel motion-based tracker, MotionTrack, centered around a learnable motion predictor.
Our experimental results demonstrate that MotionTrack yields state-of-the-art performance on datasets such as Dancetrack and SportsMOT.
arXiv Detail & Related papers (2023-06-05T04:24:11Z) - Weakly Supervised Monocular 3D Object Detection using Multi-View
Projection and Direction Consistency [78.76508318592552]
Monocular 3D object detection has become a mainstream approach in automatic driving for its easy application.
Most current methods still rely on 3D point cloud data for labeling the ground truths used in the training phase.
We propose a new weakly supervised monocular 3D objection detection method, which can train the model with only 2D labels marked on images.
arXiv Detail & Related papers (2023-03-15T15:14:00Z) - DirectTracker: 3D Multi-Object Tracking Using Direct Image Alignment and
Photometric Bundle Adjustment [41.27664827586102]
Direct methods have shown excellent performance in the applications of visual odometry and SLAM.
We propose a framework that effectively combines direct image alignment for the short-term tracking and sliding-window photometric bundle adjustment for 3D object detection.
arXiv Detail & Related papers (2022-09-29T17:40:22Z) - Image Understands Point Cloud: Weakly Supervised 3D Semantic
Segmentation via Association Learning [59.64695628433855]
We propose a novel cross-modality weakly supervised method for 3D segmentation, incorporating complementary information from unlabeled images.
Basically, we design a dual-branch network equipped with an active labeling strategy, to maximize the power of tiny parts of labels.
Our method even outperforms the state-of-the-art fully supervised competitors with less than 1% actively selected annotations.
arXiv Detail & Related papers (2022-09-16T07:59:04Z) - A Lightweight and Detector-free 3D Single Object Tracker on Point Clouds [50.54083964183614]
It is non-trivial to perform accurate target-specific detection since the point cloud of objects in raw LiDAR scans is usually sparse and incomplete.
We propose DMT, a Detector-free Motion prediction based 3D Tracking network that totally removes the usage of complicated 3D detectors.
arXiv Detail & Related papers (2022-03-08T17:49:07Z) - Beyond 3D Siamese Tracking: A Motion-Centric Paradigm for 3D Single
Object Tracking in Point Clouds [39.41305358466479]
3D single object tracking in LiDAR point clouds plays a crucial role in autonomous driving.
Current approaches all follow the Siamese paradigm based on appearance matching.
We introduce a motion-centric paradigm to handle 3D SOT from a new perspective.
arXiv Detail & Related papers (2022-03-03T14:20:10Z) - Monocular Quasi-Dense 3D Object Tracking [99.51683944057191]
A reliable and accurate 3D tracking framework is essential for predicting future locations of surrounding objects and planning the observer's actions in numerous applications such as autonomous driving.
We propose a framework that can effectively associate moving objects over time and estimate their full 3D bounding box information from a sequence of 2D images captured on a moving platform.
arXiv Detail & Related papers (2021-03-12T15:30:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.