Dynamic Semantic Occupancy Mapping using 3D Scene Flow and Closed-Form
Bayesian Inference
- URL: http://arxiv.org/abs/2108.03180v1
- Date: Fri, 6 Aug 2021 15:51:40 GMT
- Title: Dynamic Semantic Occupancy Mapping using 3D Scene Flow and Closed-Form
Bayesian Inference
- Authors: Aishwarya Unnikrishnan, Joseph Wilson, Lu Gan, Andrew Capodieci,
Paramsothy Jayakumar, Kira Barton, Maani Ghaffari
- Abstract summary: We leverage state-of-the-art semantic segmentation and 3D flow estimation using deep learning to provide measurements for map inference.
We develop a continuous (i.e., can be queried at arbitrary resolution) Bayesian model that propagates the scene with flow and infers a 3D semantic occupancy map with better performance than its static counterpart.
- Score: 3.0389083199673337
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper reports on a dynamic semantic mapping framework that incorporates
3D scene flow measurements into a closed-form Bayesian inference model.
Existence of dynamic objects in the environment cause artifacts and traces in
current mapping algorithms, leading to an inconsistent map posterior. We
leverage state-of-the-art semantic segmentation and 3D flow estimation using
deep learning to provide measurements for map inference. We develop a
continuous (i.e., can be queried at arbitrary resolution) Bayesian model that
propagates the scene with flow and infers a 3D semantic occupancy map with
better performance than its static counterpart. Experimental results using
publicly available data sets show that the proposed framework generalizes its
predecessors and improves over direct measurements from deep neural networks
consistently.
Related papers
- PMPNet: Pixel Movement Prediction Network for Monocular Depth Estimation in Dynamic Scenes [7.736445799116692]
We propose a novel method for monocular depth estimation in dynamic scenes.
We first explore the arbitrariness of object's movement trajectory in dynamic scenes theoretically.
To overcome the depth inconsistency problem around the edges, we propose a deformable support window module.
arXiv Detail & Related papers (2024-11-04T03:42:29Z) - RadOcc: Learning Cross-Modality Occupancy Knowledge through Rendering
Assisted Distillation [50.35403070279804]
3D occupancy prediction is an emerging task that aims to estimate the occupancy states and semantics of 3D scenes using multi-view images.
We propose RadOcc, a Rendering assisted distillation paradigm for 3D Occupancy prediction.
arXiv Detail & Related papers (2023-12-19T03:39:56Z) - Volumetric Semantically Consistent 3D Panoptic Mapping [77.13446499924977]
We introduce an online 2D-to-3D semantic instance mapping algorithm aimed at generating semantic 3D maps suitable for autonomous agents in unstructured environments.
It introduces novel ways of integrating semantic prediction confidence during mapping, producing semantic and instance-consistent 3D regions.
The proposed method achieves accuracy superior to the state of the art on public large-scale datasets, improving on a number of widely used metrics.
arXiv Detail & Related papers (2023-09-26T08:03:10Z) - Neural Semantic Surface Maps [52.61017226479506]
We present an automated technique for computing a map between two genus-zero shapes, which matches semantically corresponding regions to one another.
Our approach can generate semantic surface-to-surface maps, eliminating manual annotations or any 3D training data requirement.
arXiv Detail & Related papers (2023-09-09T16:21:56Z) - Incremental 3D Semantic Scene Graph Prediction from RGB Sequences [86.77318031029404]
We propose a real-time framework that incrementally builds a consistent 3D semantic scene graph of a scene given an RGB image sequence.
Our method consists of a novel incremental entity estimation pipeline and a scene graph prediction network.
The proposed network estimates 3D semantic scene graphs with iterative message passing using multi-view and geometric features extracted from the scene entities.
arXiv Detail & Related papers (2023-05-04T11:32:16Z) - Convolutional Bayesian Kernel Inference for 3D Semantic Mapping [1.7615233156139762]
We introduce a Convolutional Bayesian Kernel Inference layer which learns to perform explicit Bayesian inference.
We learn semantic-geometric probability distributions for LiDAR sensor information and incorporate semantic predictions into a global map.
We evaluate our network against state-of-the-art semantic mapping algorithms on the KITTI data set, demonstrating improved latency with comparable semantic label inference results.
arXiv Detail & Related papers (2022-09-21T21:15:12Z) - TANDEM: Tracking and Dense Mapping in Real-time using Deep Multi-view
Stereo [55.30992853477754]
We present TANDEM, a real-time monocular tracking and dense framework.
For pose estimation, TANDEM performs photometric bundle adjustment based on a sliding window of alignments.
TANDEM shows state-of-the-art real-time 3D reconstruction performance.
arXiv Detail & Related papers (2021-11-14T19:01:02Z) - SCFusion: Real-time Incremental Scene Reconstruction with Semantic
Completion [86.77318031029404]
We propose a framework that performs scene reconstruction and semantic scene completion jointly in an incremental and real-time manner.
Our framework relies on a novel neural architecture designed to process occupancy maps and leverages voxel states to accurately and efficiently fuse semantic completion with the 3D global model.
arXiv Detail & Related papers (2020-10-26T15:31:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.