GMSF: Global Matching Scene Flow
- URL: http://arxiv.org/abs/2305.17432v2
- Date: Mon, 30 Oct 2023 09:22:09 GMT
- Title: GMSF: Global Matching Scene Flow
- Authors: Yushan Zhang, Johan Edstedt, Bastian Wandt, Per-Erik Forss\'en, Maria
Magnusson, Michael Felsberg
- Abstract summary: We tackle the task of scene flow estimation from point clouds.
Given a source and a target point cloud, the objective is to estimate a translation from each point in the source point cloud to the target.
We propose a significantly simpler single-scale one-shot global matching to address the problem.
- Score: 17.077134204089536
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We tackle the task of scene flow estimation from point clouds. Given a source
and a target point cloud, the objective is to estimate a translation from each
point in the source point cloud to the target, resulting in a 3D motion vector
field. Previous dominant scene flow estimation methods require complicated
coarse-to-fine or recurrent architectures as a multi-stage refinement. In
contrast, we propose a significantly simpler single-scale one-shot global
matching to address the problem. Our key finding is that reliable feature
similarity between point pairs is essential and sufficient to estimate accurate
scene flow. We thus propose to decompose the feature extraction step via a
hybrid local-global-cross transformer architecture which is crucial to accurate
and robust feature representations. Extensive experiments show that the
proposed Global Matching Scene Flow (GMSF) sets a new state-of-the-art on
multiple scene flow estimation benchmarks. On FlyingThings3D, with the presence
of occlusion points, GMSF reduces the outlier percentage from the previous best
performance of 27.4% to 5.6%. On KITTI Scene Flow, without any fine-tuning, our
proposed method shows state-of-the-art performance. On the Waymo-Open dataset,
the proposed method outperforms previous methods by a large margin. The code is
available at https://github.com/ZhangYushan3/GMSF.
Related papers
- RMS-FlowNet++: Efficient and Robust Multi-Scale Scene Flow Estimation for Large-Scale Point Clouds [15.138542932078916]
RMS-FlowNet++ is a novel end-to-end learning-based architecture for accurate and efficient scene flow estimation.
Our architecture provides a faster prediction than state-of-the-art methods, avoids high memory requirements and enables efficient scene flow on dense point clouds of more than 250K points at once.
arXiv Detail & Related papers (2024-07-01T09:51:17Z) - Rethinking Few-shot 3D Point Cloud Semantic Segmentation [62.80639841429669]
This paper revisits few-shot 3D point cloud semantic segmentation (FS-PCS)
We focus on two significant issues in the state-of-the-art: foreground leakage and sparse point distribution.
To address these issues, we introduce a standardized FS-PCS setting, upon which a new benchmark is built.
arXiv Detail & Related papers (2024-03-01T15:14:47Z) - Grad-PU: Arbitrary-Scale Point Cloud Upsampling via Gradient Descent
with Learned Distance Functions [77.32043242988738]
We propose a new framework for accurate point cloud upsampling that supports arbitrary upsampling rates.
Our method first interpolates the low-res point cloud according to a given upsampling rate.
arXiv Detail & Related papers (2023-04-24T06:36:35Z) - PointFlowHop: Green and Interpretable Scene Flow Estimation from
Consecutive Point Clouds [49.7285297470392]
An efficient 3D scene flow estimation method called PointFlowHop is proposed in this work.
PointFlowHop takes two consecutive point clouds and determines the 3D flow vectors for every point in the first point cloud.
It decomposes the scene flow estimation task into a set of subtasks, including ego-motion compensation, object association and object-wise motion estimation.
arXiv Detail & Related papers (2023-02-27T23:06:01Z) - SCOOP: Self-Supervised Correspondence and Optimization-Based Scene Flow [25.577386156273256]
Scene flow estimation is a long-standing problem in computer vision, where the goal is to find the 3D motion of a scene from its consecutive observations.
We introduce SCOOP, a new method for scene flow estimation that can be learned on a small amount of data without employing ground-truth flow supervision.
arXiv Detail & Related papers (2022-11-25T10:52:02Z) - CloudAttention: Efficient Multi-Scale Attention Scheme For 3D Point
Cloud Learning [81.85951026033787]
We set transformers in this work and incorporate them into a hierarchical framework for shape classification and part and scene segmentation.
We also compute efficient and dynamic global cross attentions by leveraging sampling and grouping at each iteration.
The proposed hierarchical model achieves state-of-the-art shape classification in mean accuracy and yields results on par with the previous segmentation methods.
arXiv Detail & Related papers (2022-07-31T21:39:15Z) - What Matters for 3D Scene Flow Network [44.02710380584977]
3D scene flow estimation from point clouds is a low-level 3D motion perception task in computer vision.
We propose a novel all-to-all flow embedding layer with backward reliability validation during the initial scene flow estimation.
Our proposed model surpasses all existing methods by at least 38.2% on FlyingThings3D dataset and 24.7% on KITTI Scene Flow dataset for EPE3D metric.
arXiv Detail & Related papers (2022-07-19T09:27:05Z) - RMS-FlowNet: Efficient and Robust Multi-Scale Scene Flow Estimation for
Large-Scale Point Clouds [13.62166506575236]
RMS-FlowNet is a novel end-to-end learning-based architecture for accurate and efficient scene flow estimation.
We show that our model presents a competitive ability to generalize towards the real-world scenes of KITTI data set without fine-tuning.
arXiv Detail & Related papers (2022-04-01T11:02:58Z) - Stratified Transformer for 3D Point Cloud Segmentation [89.9698499437732]
Stratified Transformer is able to capture long-range contexts and demonstrates strong generalization ability and high performance.
To combat the challenges posed by irregular point arrangements, we propose first-layer point embedding to aggregate local information.
Experiments demonstrate the effectiveness and superiority of our method on S3DIS, ScanNetv2 and ShapeNetPart datasets.
arXiv Detail & Related papers (2022-03-28T05:35:16Z) - SCTN: Sparse Convolution-Transformer Network for Scene Flow Estimation [71.2856098776959]
Estimating 3D motions for point clouds is challenging, since a point cloud is unordered and its density is significantly non-uniform.
We propose a novel architecture named Sparse Convolution-Transformer Network (SCTN) that equips the sparse convolution with the transformer.
We show that the learned relation-based contextual information is rich and helpful for matching corresponding points, benefiting scene flow estimation.
arXiv Detail & Related papers (2021-05-10T15:16:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.