LODE: Locally Conditioned Eikonal Implicit Scene Completion from Sparse
LiDAR
- URL: http://arxiv.org/abs/2302.14052v1
- Date: Mon, 27 Feb 2023 18:59:58 GMT
- Title: LODE: Locally Conditioned Eikonal Implicit Scene Completion from Sparse
LiDAR
- Authors: Pengfei Li, Ruowen Zhao, Yongliang Shi, Hao Zhao, Jirui Yuan, Guyue
Zhou, Ya-Qin Zhang
- Abstract summary: Scene completion refers to obtaining dense scene representation from an incomplete perception of complex 3D scenes.
Recent advances show that implicit representation learning can be leveraged for continuous scene completion.
We propose a novel Eikonal formulation that conditions the implicit representation on localized shape priors which function as dense boundary value constraints.
- Score: 5.900616958195897
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Scene completion refers to obtaining dense scene representation from an
incomplete perception of complex 3D scenes. This helps robots detect
multi-scale obstacles and analyse object occlusions in scenarios such as
autonomous driving. Recent advances show that implicit representation learning
can be leveraged for continuous scene completion and achieved through physical
constraints like Eikonal equations. However, former Eikonal completion methods
only demonstrate results on watertight meshes at a scale of tens of meshes.
None of them are successfully done for non-watertight LiDAR point clouds of
open large scenes at a scale of thousands of scenes. In this paper, we propose
a novel Eikonal formulation that conditions the implicit representation on
localized shape priors which function as dense boundary value constraints, and
demonstrate it works on SemanticKITTI and SemanticPOSS. It can also be extended
to semantic Eikonal scene completion with only small modifications to the
network architecture. With extensive quantitative and qualitative results, we
demonstrate the benefits and drawbacks of existing Eikonal methods, which
naturally leads to the new locally conditioned formulation. Notably, we improve
IoU from 31.7% to 51.2% on SemanticKITTI and from 40.5% to 48.7% on
SemanticPOSS. We extensively ablate our methods and demonstrate that the
proposed formulation is robust to a wide spectrum of implementation
hyper-parameters. Codes and models are publicly available at
https://github.com/AIR-DISCOVER/LODE.
Related papers
- Object-level Scene Deocclusion [92.39886029550286]
We present a new self-supervised PArallel visible-to-COmplete diffusion framework, named PACO, for object-level scene deocclusion.
To train PACO, we create a large-scale dataset with 500k samples to enable self-supervised learning.
Experiments on COCOA and various real-world scenes demonstrate the superior capability of PACO for scene deocclusion, surpassing the state of the arts by a large margin.
arXiv Detail & Related papers (2024-06-11T20:34:10Z) - Mixed Diffusion for 3D Indoor Scene Synthesis [55.94569112629208]
We present MiDiffusion, a novel mixed discrete-continuous diffusion model architecture.
We represent a scene layout by a 2D floor plan and a set of objects, each defined by its category, location, size, and orientation.
Our experimental results demonstrate that MiDiffusion substantially outperforms state-of-the-art autoregressive and diffusion models in floor-conditioned 3D scene synthesis.
arXiv Detail & Related papers (2024-05-31T17:54:52Z) - Symphonize 3D Semantic Scene Completion with Contextual Instance Queries [49.604907627254434]
3D Semantic Scene Completion (SSC) has emerged as a nascent and pivotal undertaking in autonomous driving.
We present a novel paradigm termed Symphonies (Scene-from-Insts), that delves into the integration of instance queries to orchestrate 2D-to-3D reconstruction and 3D scene modeling.
arXiv Detail & Related papers (2023-06-27T17:59:46Z) - Incremental 3D Semantic Scene Graph Prediction from RGB Sequences [86.77318031029404]
We propose a real-time framework that incrementally builds a consistent 3D semantic scene graph of a scene given an RGB image sequence.
Our method consists of a novel incremental entity estimation pipeline and a scene graph prediction network.
The proposed network estimates 3D semantic scene graphs with iterative message passing using multi-view and geometric features extracted from the scene entities.
arXiv Detail & Related papers (2023-05-04T11:32:16Z) - SC-DepthV3: Robust Self-supervised Monocular Depth Estimation for
Dynamic Scenes [58.89295356901823]
Self-supervised monocular depth estimation has shown impressive results in static scenes.
It relies on the multi-view consistency assumption for training networks, however, that is violated in dynamic object regions.
We introduce an external pretrained monocular depth estimation model for generating single-image depth prior.
Our model can predict sharp and accurate depth maps, even when training from monocular videos of highly-dynamic scenes.
arXiv Detail & Related papers (2022-11-07T16:17:47Z) - CompNVS: Novel View Synthesis with Scene Completion [83.19663671794596]
We propose a generative pipeline performing on a sparse grid-based neural scene representation to complete unobserved scene parts.
We process encoded image features in 3D space with a geometry completion network and a subsequent texture inpainting network to extrapolate the missing area.
Photorealistic image sequences can be finally obtained via consistency-relevant differentiable rendering.
arXiv Detail & Related papers (2022-07-23T09:03:13Z) - Semi-supervised Implicit Scene Completion from Sparse LiDAR [11.136332180451308]
We develop a novel formulation that conditions the semi-supervised implicit function on localized shape embeddings.
It exploits the strong representation learning power of sparse convolutional networks to generate shape-aware dense feature volumes.
We demonstrate intrinsic properties of this new learning system and its usefulness in real-world road scenes.
arXiv Detail & Related papers (2021-11-29T18:50:09Z) - Semantic Scene Completion via Integrating Instances and Scene
in-the-Loop [73.11401855935726]
Semantic Scene Completion aims at reconstructing a complete 3D scene with precise voxel-wise semantics from a single-view depth or RGBD image.
We present Scene-Instance-Scene Network (textitSISNet), which takes advantages of both instance and scene level semantic information.
Our method is capable of inferring fine-grained shape details as well as nearby objects whose semantic categories are easily mixed-up.
arXiv Detail & Related papers (2021-04-08T09:50:30Z) - Semantic Scene Completion using Local Deep Implicit Functions on LiDAR
Data [4.355440821669468]
We propose a scene segmentation network based on local Deep Implicit Functions as a novel learning-based method for scene completion.
We show that this continuous representation is suitable to encode geometric and semantic properties of extensive outdoor scenes without the need for spatial discretization.
Our experiments verify that our method generates a powerful representation that can be decoded into a dense 3D description of a given scene.
arXiv Detail & Related papers (2020-11-18T07:39:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.