DoGFlow: Self-Supervised LiDAR Scene Flow via Cross-Modal Doppler Guidance
- URL: http://arxiv.org/abs/2508.18506v1
- Date: Mon, 25 Aug 2025 21:26:32 GMT
- Title: DoGFlow: Self-Supervised LiDAR Scene Flow via Cross-Modal Doppler Guidance
- Authors: Ajinkya Khoche, Qingwen Zhang, Yixi Cai, Sina Sharif Mansouri, Patric Jensfelt,
- Abstract summary: DoGFlow is a novel self-supervised framework that recovers full 3D object motions for LiDAR scene flow estimation.<n>On the challenging MAN TruckScenes dataset, DoGFlow substantially outperforms existing self-supervised methods.
- Score: 6.466382672755418
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Accurate 3D scene flow estimation is critical for autonomous systems to navigate dynamic environments safely, but creating the necessary large-scale, manually annotated datasets remains a significant bottleneck for developing robust perception models. Current self-supervised methods struggle to match the performance of fully supervised approaches, especially in challenging long-range and adverse weather scenarios, while supervised methods are not scalable due to their reliance on expensive human labeling. We introduce DoGFlow, a novel self-supervised framework that recovers full 3D object motions for LiDAR scene flow estimation without requiring any manual ground truth annotations. This paper presents our cross-modal label transfer approach, where DoGFlow computes motion pseudo-labels in real-time directly from 4D radar Doppler measurements and transfers them to the LiDAR domain using dynamic-aware association and ambiguity-resolved propagation. On the challenging MAN TruckScenes dataset, DoGFlow substantially outperforms existing self-supervised methods and improves label efficiency by enabling LiDAR backbones to achieve over 90% of fully supervised performance with only 10% of the ground truth data. For more details, please visit https://ajinkyakhoche.github.io/DogFlow/
Related papers
- DriveFlow: Rectified Flow Adaptation for Robust 3D Object Detection in Autonomous Driving [85.14946767994932]
DriveFlow is a Rectified Flow Adaptation method for training data enhancement in autonomous driving.<n>It incorporates a high-frequency alignment loss for foreground to maintain precise 3D object geometry.<n>It also conducts dual-frequency optimization for background, balancing editing flexibility and semantic consistency.
arXiv Detail & Related papers (2025-11-24T03:12:43Z) - UniFlow: Towards Zero-Shot LiDAR Scene Flow for Autonomous Vehicles via Cross-Domain Generalization [46.6187296172395]
We learn general motion priors that transfer to diverse and unseen LiDAR sensors.<n>We propose UniFlow, a family of feedforward models that unifies and trains on multiple large-scale LiDAR scene flow datasets.<n>UniFlow achieves state-of-the-art accuracy on unseen datasets like TruckScenes, outperforming prior TruckScenes-specific models by 30.1%.
arXiv Detail & Related papers (2025-11-23T02:51:42Z) - Learning to Track Any Points from Human Motion [55.831218129679144]
We propose an automated pipeline to generate pseudo-labeled training data for point tracking.<n>A point tracking model trained on AnthroTAP achieves annotated state-of-the-art performance on the TAP-Vid benchmark.
arXiv Detail & Related papers (2025-07-08T17:59:58Z) - SuperFlow++: Enhanced Spatiotemporal Consistency for Cross-Modal Data Pretraining [62.433137130087445]
SuperFlow++ is a novel framework that integrates pretraining and downstream tasks using consecutive camera pairs.<n>We show that SuperFlow++ outperforms state-of-the-art methods across diverse tasks and driving conditions.<n>With strong generalizability and computational efficiency, SuperFlow++ establishes a new benchmark for data-efficient LiDAR-based perception in autonomous driving.
arXiv Detail & Related papers (2025-03-25T17:59:57Z) - SeFlow: A Self-Supervised Scene Flow Method in Autonomous Driving [18.88208422580103]
Scene flow estimation predicts the 3D motion at each point in successive LiDAR scans.
Current state-of-the-art methods require annotated data to train scene flow networks.
We propose SeFlow, a self-supervised method that integrates efficient dynamic classification into a learning-based scene flow pipeline.
arXiv Detail & Related papers (2024-07-01T18:22:54Z) - 3DSFLabelling: Boosting 3D Scene Flow Estimation by Pseudo
Auto-labelling [21.726386822643995]
We present a novel approach to generate a large number of 3D scene flow pseudo labels for real-world LiDAR point clouds.
Specifically, we employ the assumption of rigid body motion to simulate potential object-level rigid movements in autonomous driving scenarios.
By perfectly synthesizing target point clouds based on augmented motion parameters, we easily obtain lots of 3D scene flow labels in point clouds highly consistent with real scenarios.
arXiv Detail & Related papers (2024-02-28T08:12:31Z) - Unsupervised Domain Adaptation for Self-Driving from Past Traversal
Features [69.47588461101925]
We propose a method to adapt 3D object detectors to new driving environments.
Our approach enhances LiDAR-based detection models using spatial quantized historical features.
Experiments on real-world datasets demonstrate significant improvements.
arXiv Detail & Related papers (2023-09-21T15:00:31Z) - ZeroFlow: Scalable Scene Flow via Distillation [66.70820145266029]
Scene flow estimation is the task of describing the 3D motion field between temporally successive point clouds.
State-of-the-art methods use strong priors and test-time optimization techniques, but require on the order of tens of seconds to process full-size point clouds.
We propose Scene Flow via Distillation, a simple, scalable distillation framework that uses a label-free optimization method to produce pseudo-labels to supervise a feedforward model.
arXiv Detail & Related papers (2023-05-17T17:56:59Z) - An Effective Motion-Centric Paradigm for 3D Single Object Tracking in
Point Clouds [50.19288542498838]
3D single object tracking in LiDAR point clouds (LiDAR SOT) plays a crucial role in autonomous driving.
Current approaches all follow the Siamese paradigm based on appearance matching.
We introduce a motion-centric paradigm to handle LiDAR SOT from a new perspective.
arXiv Detail & Related papers (2023-03-21T17:28:44Z) - FAST3D: Flow-Aware Self-Training for 3D Object Detectors [12.511087244102036]
State-of-the-art self-training approaches mostly ignore the temporal nature of autonomous driving data.
We propose a flow-aware self-training method that enables unsupervised domain adaptation for 3D object detectors on continuous LiDAR point clouds.
Our results show a significant improvement over the state-of-the-art, without any prior target domain knowledge.
arXiv Detail & Related papers (2021-10-18T14:32:05Z) - Weakly Supervised Learning of Rigid 3D Scene Flow [81.37165332656612]
We propose a data-driven scene flow estimation algorithm exploiting the observation that many 3D scenes can be explained by a collection of agents moving as rigid bodies.
We showcase the effectiveness and generalization capacity of our method on four different autonomous driving datasets.
arXiv Detail & Related papers (2021-02-17T18:58:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.