Differentiable Rendering for Pose Estimation in Proximity Operations
- URL: http://arxiv.org/abs/2212.12668v1
- Date: Sat, 24 Dec 2022 06:12:16 GMT
- Title: Differentiable Rendering for Pose Estimation in Proximity Operations
- Authors: Ramchander Rao Bhaskara and Roshan Thomas Eapen and Manoranjan Majji
- Abstract summary: Differentiable rendering aims to compute the derivative of the image rendering function with respect to the rendering parameters.
This paper presents a novel algorithm for 6-DoF pose estimation using a differentiable rendering pipeline.
- Score: 4.282159812965446
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Differentiable rendering aims to compute the derivative of the image
rendering function with respect to the rendering parameters. This paper
presents a novel algorithm for 6-DoF pose estimation through gradient-based
optimization using a differentiable rendering pipeline. We emphasize two key
contributions: (1) instead of solving the conventional 2D to 3D correspondence
problem and computing reprojection errors, images (rendered using the 3D model)
are compared only in the 2D feature space via sparse 2D feature
correspondences. (2) Instead of an analytical image formation model, we compute
an approximate local gradient of the rendering process through online learning.
The learning data consists of image features extracted from multi-viewpoint
renders at small perturbations in the pose neighborhood. The gradients are
propagated through the rendering pipeline for the 6-DoF pose estimation using
nonlinear least squares. This gradient-based optimization regresses directly
upon the pose parameters by aligning the 3D model to reproduce a reference
image shape. Using representative experiments, we demonstrate the application
of our approach to pose estimation in proximity operations.
Related papers
- GPS-Gaussian+: Generalizable Pixel-wise 3D Gaussian Splatting for Real-Time Human-Scene Rendering from Sparse Views [67.34073368933814]
We propose a generalizable Gaussian Splatting approach for high-resolution image rendering under a sparse-view camera setting.
We train our Gaussian parameter regression module on human-only data or human-scene data, jointly with a depth estimation module to lift 2D parameter maps to 3D space.
Experiments on several datasets demonstrate that our method outperforms state-of-the-art methods while achieving an exceeding rendering speed.
arXiv Detail & Related papers (2024-11-18T08:18:44Z) - Towards Human-Level 3D Relative Pose Estimation: Generalizable, Training-Free, with Single Reference [62.99706119370521]
Humans can easily deduce the relative pose of an unseen object, without label/training, given only a single query-reference image pair.
We propose a novel 3D generalizable relative pose estimation method by elaborating (i) with a 2.5D shape from an RGB-D reference, (ii) with an off-the-shelf differentiable, and (iii) with semantic cues from a pretrained model like DINOv2.
arXiv Detail & Related papers (2024-06-26T16:01:10Z) - Personalized 3D Human Pose and Shape Refinement [19.082329060985455]
regression-based methods have dominated the field of 3D human pose and shape estimation.
We propose to construct dense correspondences between initial human model estimates and the corresponding images.
We show that our approach not only consistently leads to better image-model alignment, but also to improved 3D accuracy.
arXiv Detail & Related papers (2024-03-18T10:13:53Z) - GAN2X: Non-Lambertian Inverse Rendering of Image GANs [85.76426471872855]
We present GAN2X, a new method for unsupervised inverse rendering that only uses unpaired images for training.
Unlike previous Shape-from-GAN approaches that mainly focus on 3D shapes, we take the first attempt to also recover non-Lambertian material properties by exploiting the pseudo paired data generated by a GAN.
Experiments demonstrate that GAN2X can accurately decompose 2D images to 3D shape, albedo, and specular properties for different object categories, and achieves the state-of-the-art performance for unsupervised single-view 3D face reconstruction.
arXiv Detail & Related papers (2022-06-18T16:58:49Z) - A Shading-Guided Generative Implicit Model for Shape-Accurate 3D-Aware
Image Synthesis [163.96778522283967]
We propose a shading-guided generative implicit model that is able to learn a starkly improved shape representation.
An accurate 3D shape should also yield a realistic rendering under different lighting conditions.
Our experiments on multiple datasets show that the proposed approach achieves photorealistic 3D-aware image synthesis.
arXiv Detail & Related papers (2021-10-29T10:53:12Z) - Differentiable Rendering with Perturbed Optimizers [85.66675707599782]
Reasoning about 3D scenes from their 2D image projections is one of the core problems in computer vision.
Our work highlights the link between some well-known differentiable formulations and randomly smoothed renderings.
We apply our method to 3D scene reconstruction and demonstrate its advantages on the tasks of 6D pose estimation and 3D mesh reconstruction.
arXiv Detail & Related papers (2021-10-18T08:56:23Z) - Using Adaptive Gradient for Texture Learning in Single-View 3D
Reconstruction [0.0]
Learning-based approaches for 3D model reconstruction have attracted attention owing to its modern applications.
We present a novel sampling algorithm by optimizing the gradient of predicted coordinates based on the variance on the sampling image.
We also adopt Frechet Inception Distance (FID) to form a loss function in learning, which helps bridging the gap between rendered images and input images.
arXiv Detail & Related papers (2021-04-29T07:52:54Z) - Geometric Correspondence Fields: Learned Differentiable Rendering for 3D
Pose Refinement in the Wild [96.09941542587865]
We present a novel 3D pose refinement approach based on differentiable rendering for objects of arbitrary categories in the wild.
In this way, we precisely align 3D models to objects in RGB images which results in significantly improved 3D pose estimates.
We evaluate our approach on the challenging Pix3D dataset and achieve up to 55% relative improvement compared to state-of-the-art refinement methods in multiple metrics.
arXiv Detail & Related papers (2020-07-17T12:34:38Z) - Shape from Projections via Differentiable Forward Projector for Computed
Tomography [4.304380400377787]
We propose a differentiable forward model for 3D meshes that bridge the gap between the forward model for 3D surfaces and optimization.
We use the proposed forward model to reconstruct 3D shapes directly from projections.
Experimental results for single-object problems show that the proposed method outperforms traditional voxel-based methods on noisy simulated data.
arXiv Detail & Related papers (2020-06-29T15:33:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.