Graph-Based Depth Denoising & Dequantization for Point Cloud Enhancement
- URL: http://arxiv.org/abs/2111.04946v1
- Date: Tue, 9 Nov 2021 04:17:35 GMT
- Title: Graph-Based Depth Denoising & Dequantization for Point Cloud Enhancement
- Authors: Xue Zhang, Gene Cheung, Jiahao Pang, Yash Sanghvi, Abhiram
Gnanasambandam, Stanley H. Chan
- Abstract summary: A 3D point cloud is typically constructed from depth measurements acquired by sensors at one or more viewpoints.
Previous works denoise a point cloud textita posteriori after projecting the imperfect depth data onto 3D space.
We enhance depth measurements directly on the sensed images textita priori, before synthesizing a 3D point cloud.
- Score: 47.61748619439693
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A 3D point cloud is typically constructed from depth measurements acquired by
sensors at one or more viewpoints. The measurements suffer from both
quantization and noise corruption. To improve quality, previous works denoise a
point cloud \textit{a posteriori} after projecting the imperfect depth data
onto 3D space. Instead, we enhance depth measurements directly on the sensed
images \textit{a priori}, before synthesizing a 3D point cloud. By enhancing
near the physical sensing process, we tailor our optimization to our depth
formation model before subsequent processing steps that obscure measurement
errors. Specifically, we model depth formation as a combined process of
signal-dependent noise addition and non-uniform log-based quantization. The
designed model is validated (with parameters fitted) using collected empirical
data from an actual depth sensor. To enhance each pixel row in a depth image,
we first encode intra-view similarities between available row pixels as edge
weights via feature graph learning. We next establish inter-view similarities
with another rectified depth image via viewpoint mapping and sparse linear
interpolation. This leads to a maximum a posteriori (MAP) graph filtering
objective that is convex and differentiable. We optimize the objective
efficiently using accelerated gradient descent (AGD), where the optimal step
size is approximated via Gershgorin circle theorem (GCT). Experiments show that
our method significantly outperformed recent point cloud denoising schemes and
state-of-the-art image denoising schemes, in two established point cloud
quality metrics.
Related papers
- PointRegGPT: Boosting 3D Point Cloud Registration using Generative Point-Cloud Pairs for Training [90.06520673092702]
We present PointRegGPT, boosting 3D point cloud registration using generative point-cloud pairs for training.
To our knowledge, this is the first generative approach that explores realistic data generation for indoor point cloud registration.
arXiv Detail & Related papers (2024-07-19T06:29:57Z) - A Construct-Optimize Approach to Sparse View Synthesis without Camera Pose [44.13819148680788]
We develop a novel construct-and-optimize method for sparse view synthesis without camera poses.
Specifically, we construct a solution by using monocular depth and projecting pixels back into the 3D world.
We demonstrate results on the Tanks and Temples and Static Hikes datasets with as few as three widely-spaced views.
arXiv Detail & Related papers (2024-05-06T17:36:44Z) - Depth-Regularized Optimization for 3D Gaussian Splatting in Few-Shot
Images [47.14713579719103]
We introduce a dense depth map as a geometry guide to mitigate overfitting.
The adjusted depth aids in the color-based optimization of 3D Gaussian splatting.
We verify the proposed method on the NeRF-LLFF dataset with varying numbers of few images.
arXiv Detail & Related papers (2023-11-22T13:53:04Z) - Diff-DOPE: Differentiable Deep Object Pose Estimation [29.703385848843414]
We introduce Diff-DOPE, a 6-DoF pose refiner that takes as input an image, a 3D textured model of an object, and an initial pose of the object.
The method uses differentiable rendering to update the object pose to minimize the visual error between the image and the projection of the model.
We show that this simple, yet effective, idea is able to achieve state-of-the-art results on pose estimation datasets.
arXiv Detail & Related papers (2023-09-30T18:52:57Z) - $PC^2$: Projection-Conditioned Point Cloud Diffusion for Single-Image 3D
Reconstruction [97.06927852165464]
Reconstructing the 3D shape of an object from a single RGB image is a long-standing and highly challenging problem in computer vision.
We propose a novel method for single-image 3D reconstruction which generates a sparse point cloud via a conditional denoising diffusion process.
arXiv Detail & Related papers (2023-02-21T13:37:07Z) - GraphCSPN: Geometry-Aware Depth Completion via Dynamic GCNs [49.55919802779889]
We propose a Graph Convolution based Spatial Propagation Network (GraphCSPN) as a general approach for depth completion.
In this work, we leverage convolution neural networks as well as graph neural networks in a complementary way for geometric representation learning.
Our method achieves the state-of-the-art performance, especially when compared in the case of using only a few propagation steps.
arXiv Detail & Related papers (2022-10-19T17:56:03Z) - Virtual Normal: Enforcing Geometric Constraints for Accurate and Robust
Depth Prediction [87.08227378010874]
We show the importance of the high-order 3D geometric constraints for depth prediction.
By designing a loss term that enforces a simple geometric constraint, we significantly improve the accuracy and robustness of monocular depth estimation.
We show state-of-the-art results of learning metric depth on NYU Depth-V2 and KITTI.
arXiv Detail & Related papers (2021-03-07T00:08:21Z) - Depth Completion Using a View-constrained Deep Prior [73.21559000917554]
Recent work has shown that the structure of convolutional neural networks (CNNs) induces a strong prior that favors natural images.
This prior, known as a deep image prior (DIP), is an effective regularizer in inverse problems such as image denoising and inpainting.
We extend the concept of the DIP to depth images. Given color images and noisy and incomplete target depth maps, we reconstruct a depth map restored by virtue of using the CNN network structure as a prior.
arXiv Detail & Related papers (2020-01-21T21:56:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.