DeFlow: Decoder of Scene Flow Network in Autonomous Driving
- URL: http://arxiv.org/abs/2401.16122v1
- Date: Mon, 29 Jan 2024 12:47:55 GMT
- Title: DeFlow: Decoder of Scene Flow Network in Autonomous Driving
- Authors: Qingwen Zhang, Yi Yang, Heng Fang, Ruoyu Geng, Patric Jensfelt
- Abstract summary: Scene flow estimation determines a scene's 3D motion field, by predicting the motion of points in the scene.
Many networks with large-scale point clouds as input use voxelization to create a pseudo-image for real-time running.
Our paper introduces DeFlow which enables a transition from voxel-based features to point features using Gated Recurrent Unit (GRU) refinement.
- Score: 19.486167661795797
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Scene flow estimation determines a scene's 3D motion field, by predicting the
motion of points in the scene, especially for aiding tasks in autonomous
driving. Many networks with large-scale point clouds as input use voxelization
to create a pseudo-image for real-time running. However, the voxelization
process often results in the loss of point-specific features. This gives rise
to a challenge in recovering those features for scene flow tasks. Our paper
introduces DeFlow which enables a transition from voxel-based features to point
features using Gated Recurrent Unit (GRU) refinement. To further enhance scene
flow estimation performance, we formulate a novel loss function that accounts
for the data imbalance between static and dynamic points. Evaluations on the
Argoverse 2 scene flow task reveal that DeFlow achieves state-of-the-art
results on large-scale point cloud data, demonstrating that our network has
better performance and efficiency compared to others. The code is open-sourced
at https://github.com/KTH-RPL/deflow.
Related papers
- Neural Eulerian Scene Flow Fields [59.57980592109722]
EulerFlow works out-of-the-box without tuning across multiple domains.
It exhibits emergent 3D point tracking behavior by solving its estimated ODE over long-time horizons.
It outperforms all prior art on the Argoverse 2 2024 Scene Flow Challenge.
arXiv Detail & Related papers (2024-10-02T20:56:45Z) - SeFlow: A Self-Supervised Scene Flow Method in Autonomous Driving [18.88208422580103]
Scene flow estimation predicts the 3D motion at each point in successive LiDAR scans.
Current state-of-the-art methods require annotated data to train scene flow networks.
We propose SeFlow, a self-supervised method that integrates efficient dynamic classification into a learning-based scene flow pipeline.
arXiv Detail & Related papers (2024-07-01T18:22:54Z) - PointFlowHop: Green and Interpretable Scene Flow Estimation from
Consecutive Point Clouds [49.7285297470392]
An efficient 3D scene flow estimation method called PointFlowHop is proposed in this work.
PointFlowHop takes two consecutive point clouds and determines the 3D flow vectors for every point in the first point cloud.
It decomposes the scene flow estimation task into a set of subtasks, including ego-motion compensation, object association and object-wise motion estimation.
arXiv Detail & Related papers (2023-02-27T23:06:01Z) - SCOOP: Self-Supervised Correspondence and Optimization-Based Scene Flow [25.577386156273256]
Scene flow estimation is a long-standing problem in computer vision, where the goal is to find the 3D motion of a scene from its consecutive observations.
We introduce SCOOP, a new method for scene flow estimation that can be learned on a small amount of data without employing ground-truth flow supervision.
arXiv Detail & Related papers (2022-11-25T10:52:02Z) - What Matters for 3D Scene Flow Network [44.02710380584977]
3D scene flow estimation from point clouds is a low-level 3D motion perception task in computer vision.
We propose a novel all-to-all flow embedding layer with backward reliability validation during the initial scene flow estimation.
Our proposed model surpasses all existing methods by at least 38.2% on FlyingThings3D dataset and 24.7% on KITTI Scene Flow dataset for EPE3D metric.
arXiv Detail & Related papers (2022-07-19T09:27:05Z) - Real-time Object Detection for Streaming Perception [84.2559631820007]
Streaming perception is proposed to jointly evaluate the latency and accuracy into a single metric for video online perception.
We build a simple and effective framework for streaming perception.
Our method achieves competitive performance on Argoverse-HD dataset and improves the AP by 4.9% compared to the strong baseline.
arXiv Detail & Related papers (2022-03-23T11:33:27Z) - SCTN: Sparse Convolution-Transformer Network for Scene Flow Estimation [71.2856098776959]
Estimating 3D motions for point clouds is challenging, since a point cloud is unordered and its density is significantly non-uniform.
We propose a novel architecture named Sparse Convolution-Transformer Network (SCTN) that equips the sparse convolution with the transformer.
We show that the learned relation-based contextual information is rich and helpful for matching corresponding points, benefiting scene flow estimation.
arXiv Detail & Related papers (2021-05-10T15:16:14Z) - Scene Flow from Point Clouds with or without Learning [47.03163552693887]
Scene flow is the three-dimensional (3D) motion field of a scene.
Current learning-based approaches seek to estimate the scene flow directly from point clouds.
We present a simple and interpretable objective function to recover the scene flow from point clouds.
arXiv Detail & Related papers (2020-10-31T17:24:48Z) - Hierarchical Attention Learning of Scene Flow in 3D Point Clouds [28.59260783047209]
This paper studies the problem of scene flow estimation from two consecutive 3D point clouds.
A novel hierarchical neural network with double attention is proposed for learning the correlation of point features in adjacent frames.
Experiments show that the proposed network outperforms the state-of-the-art performance of 3D scene flow estimation.
arXiv Detail & Related papers (2020-10-12T14:56:08Z) - Feature Flow: In-network Feature Flow Estimation for Video Object
Detection [56.80974623192569]
Optical flow is widely used in computer vision tasks to provide pixel-level motion information.
A common approach is to:forward optical flow to a neural network and fine-tune this network on the task dataset.
We propose a novel network (IFF-Net) with an textbfIn-network textbfFeature textbfFlow estimation module for video object detection.
arXiv Detail & Related papers (2020-09-21T07:55:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.