Self-Supervised Bird's Eye View Motion Prediction with Cross-Modality
Signals
- URL: http://arxiv.org/abs/2401.11499v1
- Date: Sun, 21 Jan 2024 14:09:49 GMT
- Title: Self-Supervised Bird's Eye View Motion Prediction with Cross-Modality
Signals
- Authors: Shaoheng Fang, Zuhong Liu, Mingyu Wang, Chenxin Xu, Yiqi Zhong, Siheng
Chen
- Abstract summary: Learning the dense bird's eye view (BEV) motion flow in a self-supervised manner is an emerging research for robotics and autonomous driving.
Current self-supervised methods mainly rely on point correspondences between point clouds.
We introduce a novel cross-modality self-supervised training framework that effectively addresses these issues by leveraging multi-modality data.
- Score: 38.20643428486824
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning the dense bird's eye view (BEV) motion flow in a self-supervised
manner is an emerging research for robotics and autonomous driving. Current
self-supervised methods mainly rely on point correspondences between point
clouds, which may introduce the problems of fake flow and inconsistency,
hindering the model's ability to learn accurate and realistic motion. In this
paper, we introduce a novel cross-modality self-supervised training framework
that effectively addresses these issues by leveraging multi-modality data to
obtain supervision signals. We design three innovative supervision signals to
preserve the inherent properties of scene motion, including the masked Chamfer
distance loss, the piecewise rigidity loss, and the temporal consistency loss.
Through extensive experiments, we demonstrate that our proposed self-supervised
framework outperforms all previous self-supervision methods for the motion
prediction task.
Related papers
- OOSTraj: Out-of-Sight Trajectory Prediction With Vision-Positioning Denoising [49.86409475232849]
Trajectory prediction is fundamental in computer vision and autonomous driving.
Existing approaches in this field often assume precise and complete observational data.
We present a novel method for out-of-sight trajectory prediction that leverages a vision-positioning technique.
arXiv Detail & Related papers (2024-04-02T18:30:29Z) - Self-Supervised Class-Agnostic Motion Prediction with Spatial and Temporal Consistency Regularizations [53.797896854533384]
Class-agnostic motion prediction methods directly predict the motion of the entire point cloud.
While most existing methods rely on fully-supervised learning, the manual labeling of point cloud data is laborious and time-consuming.
We introduce three simple spatial and temporal regularization losses, which facilitate the self-supervised training process effectively.
arXiv Detail & Related papers (2024-03-20T02:58:45Z) - EMR-MSF: Self-Supervised Recurrent Monocular Scene Flow Exploiting
Ego-Motion Rigidity [13.02735046166494]
Self-supervised monocular scene flow estimation has received increasing attention for its simple and economical sensor setup.
We propose a superior model named EMR-MSF by borrowing the advantages of network architecture design under the scope of supervised learning.
On the KITTI scene flow benchmark, our approach improves the SF-all metric of the state-of-the-art self-supervised monocular method by 44%.
arXiv Detail & Related papers (2023-09-04T00:30:06Z) - Smooth-Trajectron++: Augmenting the Trajectron++ behaviour prediction
model with smooth attention [0.0]
This work investigates the state-of-the-art trajectory forecasting model Trajectron++ which we enhance by incorporating a smoothing term in its attention module.
This attention mechanism mimics human attention inspired by cognitive science research indicating limits to attention switching.
We evaluate the performance of the resulting Smooth-Trajectron++ model and compare it to the original model on various benchmarks.
arXiv Detail & Related papers (2023-05-31T09:19:55Z) - Model-Based Reinforcement Learning with Isolated Imaginations [61.67183143982074]
We propose Iso-Dream++, a model-based reinforcement learning approach.
We perform policy optimization based on the decoupled latent imaginations.
This enables long-horizon visuomotor control tasks to benefit from isolating mixed dynamics sources in the wild.
arXiv Detail & Related papers (2023-03-27T02:55:56Z) - Motion Inspired Unsupervised Perception and Prediction in Autonomous
Driving [29.731790562352344]
This paper pioneers a novel and challenging direction, i.e., training perception and prediction models to understand open-set moving objects.
Our proposed framework uses self-learned flow to trigger an automated meta labeling pipeline to achieve automatic supervision.
We show that our approach generates highly promising results in open-set 3D detection and trajectory prediction.
arXiv Detail & Related papers (2022-10-14T18:55:44Z) - MotionHint: Self-Supervised Monocular Visual Odometry with Motion
Constraints [70.76761166614511]
We present a novel self-supervised algorithm named MotionHint for monocular visual odometry (VO)
Our MotionHint algorithm can be easily applied to existing open-sourced state-of-the-art SSM-VO systems.
arXiv Detail & Related papers (2021-09-14T15:35:08Z) - Self-Supervised Pillar Motion Learning for Autonomous Driving [10.921208239968827]
We propose a learning framework that leverages free supervisory signals from point clouds and paired camera images to estimate motion purely via self-supervision.
Our model involves a point cloud based structural consistency augmented with probabilistic motion masking as well as a cross-sensor motion regularization to realize the desired self-supervision.
arXiv Detail & Related papers (2021-04-18T02:32:08Z) - Self-supervised Video Object Segmentation [76.83567326586162]
The objective of this paper is self-supervised representation learning, with the goal of solving semi-supervised video object segmentation (a.k.a. dense tracking)
We make the following contributions: (i) we propose to improve the existing self-supervised approach, with a simple, yet more effective memory mechanism for long-term correspondence matching; (ii) by augmenting the self-supervised approach with an online adaptation module, our method successfully alleviates tracker drifts caused by spatial-temporal discontinuity; (iv) we demonstrate state-of-the-art results among the self-supervised approaches on DAVIS-2017 and YouTube
arXiv Detail & Related papers (2020-06-22T17:55:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.