Multi-Object Self-Supervised Depth Denoising
- URL: http://arxiv.org/abs/2305.05778v1
- Date: Tue, 9 May 2023 21:48:44 GMT
- Title: Multi-Object Self-Supervised Depth Denoising
- Authors: Claudius Kienle and David Petri
- Abstract summary: Small and compact depth cameras are often not sufficient for precise tracking in and perception of the robot's working space.
We present a self-supervised multi-object depth denoising pipeline, that uses depth maps of higher-quality sensors as close-to-ground-truth supervisory signals to denoise depth maps coming from a lower-quality sensor.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Depth cameras are frequently used in robotic manipulation, e.g. for visual
servoing. The quality of small and compact depth cameras is though often not
sufficient for depth reconstruction, which is required for precise tracking in
and perception of the robot's working space. Based on the work of Shabanov et
al. (2021), in this work, we present a self-supervised multi-object depth
denoising pipeline, that uses depth maps of higher-quality sensors as
close-to-ground-truth supervisory signals to denoise depth maps coming from a
lower-quality sensor. We display a computationally efficient way to align sets
of two frame pairs in space and retrieve a frame-based multi-object mask, in
order to receive a clean labeled dataset to train a denoising neural network
on. The implementation of our presented work can be found at
https://github.com/alr-internship/self-supervised-depth-denoising.
Related papers
- Neural Implicit Dense Semantic SLAM [83.04331351572277]
We propose a novel RGBD vSLAM algorithm that learns a memory-efficient, dense 3D geometry, and semantic segmentation of an indoor scene in an online manner.
Our pipeline combines classical 3D vision-based tracking and loop closing with neural fields-based mapping.
Our proposed algorithm can greatly enhance scene perception and assist with a range of robot control problems.
arXiv Detail & Related papers (2023-04-27T23:03:52Z) - Multi-Camera Collaborative Depth Prediction via Consistent Structure
Estimation [75.99435808648784]
We propose a novel multi-camera collaborative depth prediction method.
It does not require large overlapping areas while maintaining structure consistency between cameras.
Experimental results on DDAD and NuScenes datasets demonstrate the superior performance of our method.
arXiv Detail & Related papers (2022-10-05T03:44:34Z) - Unsupervised Depth Completion with Calibrated Backprojection Layers [79.35651668390496]
We propose a deep neural network architecture to infer dense depth from an image and a sparse point cloud.
It is trained using a video stream and corresponding synchronized sparse point cloud, as obtained from a LIDAR or other range sensor, along with the intrinsic calibration parameters of the camera.
At inference time, the calibration of the camera, which can be different from the one used for training, is fed as an input to the network along with the sparse point cloud and a single image.
arXiv Detail & Related papers (2021-08-24T05:41:59Z) - DnD: Dense Depth Estimation in Crowded Dynamic Indoor Scenes [68.38952377590499]
We present a novel approach for estimating depth from a monocular camera as it moves through complex indoor environments.
Our approach predicts absolute scale depth maps over the entire scene consisting of a static background and multiple moving people.
arXiv Detail & Related papers (2021-08-12T09:12:39Z) - Sparse Auxiliary Networks for Unified Monocular Depth Prediction and
Completion [56.85837052421469]
Estimating scene geometry from data obtained with cost-effective sensors is key for robots and self-driving cars.
In this paper, we study the problem of predicting dense depth from a single RGB image with optional sparse measurements from low-cost active depth sensors.
We introduce Sparse Networks (SANs), a new module enabling monodepth networks to perform both the tasks of depth prediction and completion.
arXiv Detail & Related papers (2021-03-30T21:22:26Z) - Multi-Modal Depth Estimation Using Convolutional Neural Networks [0.8701566919381223]
This paper addresses the problem of dense depth predictions from sparse distance sensor data and a single camera image on challenging weather conditions.
It explores the significance of different sensor modalities such as camera, Radar, and Lidar for estimating depth by applying Deep Learning approaches.
arXiv Detail & Related papers (2020-12-17T15:31:49Z) - Robust Consistent Video Depth Estimation [65.53308117778361]
We present an algorithm for estimating consistent dense depth maps and camera poses from a monocular video.
Our algorithm combines two complementary techniques: (1) flexible deformation-splines for low-frequency large-scale alignment and (2) geometry-aware depth filtering for high-frequency alignment of fine depth details.
In contrast to prior approaches, our method does not require camera poses as input and achieves robust reconstruction for challenging hand-held cell phone captures containing a significant amount of noise, shake, motion blur, and rolling shutter deformations.
arXiv Detail & Related papers (2020-12-10T18:59:48Z) - Self-supervised Depth Denoising Using Lower- and Higher-quality RGB-D
sensors [8.34403807284064]
We propose a self-supervised depth denoising approach to denoise and refine depth coming from a low quality sensor.
We record simultaneous RGB-D sequences with unzynchronized lower- and higher-quality cameras and solve a challenging problem of aligning sequences both temporally and spatially.
We then learn a deep neural network to denoise the lower-quality depth using the matched higher-quality data as a source of supervision signal.
arXiv Detail & Related papers (2020-09-10T11:18:11Z) - Self-Attention Dense Depth Estimation Network for Unrectified Video
Sequences [6.821598757786515]
LiDAR and radar sensors are the hardware solution for real-time depth estimation.
Deep learning based self-supervised depth estimation methods have shown promising results.
We propose a self-attention based depth and ego-motion network for unrectified images.
arXiv Detail & Related papers (2020-05-28T21:53:53Z) - Depth Map Estimation of Dynamic Scenes Using Prior Depth Information [14.03714478207425]
We propose an algorithm that estimates depth maps using concurrently collected images and a previously measured depth map for dynamic scenes.
Our goal is to balance the acquisition of depth between the active depth sensor and computation, without incurring a large computational cost.
Our approach can obtain dense depth maps at up to real-time (30 FPS) on a standard laptop computer, which is orders of magnitude faster than similar approaches.
arXiv Detail & Related papers (2020-02-02T01:04:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.