Geometric Correspondence Fields: Learned Differentiable Rendering for 3D
Pose Refinement in the Wild
- URL: http://arxiv.org/abs/2007.08939v1
- Date: Fri, 17 Jul 2020 12:34:38 GMT
- Title: Geometric Correspondence Fields: Learned Differentiable Rendering for 3D
Pose Refinement in the Wild
- Authors: Alexander Grabner, Yaming Wang, Peizhao Zhang, Peihong Guo, Tong Xiao,
Peter Vajda, Peter M. Roth, Vincent Lepetit
- Abstract summary: We present a novel 3D pose refinement approach based on differentiable rendering for objects of arbitrary categories in the wild.
In this way, we precisely align 3D models to objects in RGB images which results in significantly improved 3D pose estimates.
We evaluate our approach on the challenging Pix3D dataset and achieve up to 55% relative improvement compared to state-of-the-art refinement methods in multiple metrics.
- Score: 96.09941542587865
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a novel 3D pose refinement approach based on differentiable
rendering for objects of arbitrary categories in the wild. In contrast to
previous methods, we make two main contributions: First, instead of comparing
real-world images and synthetic renderings in the RGB or mask space, we compare
them in a feature space optimized for 3D pose refinement. Second, we introduce
a novel differentiable renderer that learns to approximate the rasterization
backward pass from data instead of relying on a hand-crafted algorithm. For
this purpose, we predict deep cross-domain correspondences between RGB images
and 3D model renderings in the form of what we call geometric correspondence
fields. These correspondence fields serve as pixel-level gradients which are
analytically propagated backward through the rendering pipeline to perform a
gradient-based optimization directly on the 3D pose. In this way, we precisely
align 3D models to objects in RGB images which results in significantly
improved 3D pose estimates. We evaluate our approach on the challenging Pix3D
dataset and achieve up to 55% relative improvement compared to state-of-the-art
refinement methods in multiple metrics.
Related papers
- Enhancing Single Image to 3D Generation using Gaussian Splatting and Hybrid Diffusion Priors [17.544733016978928]
3D object generation from a single image involves estimating the full 3D geometry and texture of unseen views from an unposed RGB image captured in the wild.
Recent advancements in 3D object generation have introduced techniques that reconstruct an object's 3D shape and texture.
We propose bridging the gap between 2D and 3D diffusion models to address this limitation.
arXiv Detail & Related papers (2024-10-12T10:14:11Z) - Towards Human-Level 3D Relative Pose Estimation: Generalizable, Training-Free, with Single Reference [62.99706119370521]
Humans can easily deduce the relative pose of an unseen object, without label/training, given only a single query-reference image pair.
We propose a novel 3D generalizable relative pose estimation method by elaborating (i) with a 2.5D shape from an RGB-D reference, (ii) with an off-the-shelf differentiable, and (iii) with semantic cues from a pretrained model like DINOv2.
arXiv Detail & Related papers (2024-06-26T16:01:10Z) - Sketch3D: Style-Consistent Guidance for Sketch-to-3D Generation [55.73399465968594]
This paper proposes a novel generation paradigm Sketch3D to generate realistic 3D assets with shape aligned with the input sketch and color matching the textual description.
Three strategies are designed to optimize 3D Gaussians, i.e., structural optimization via a distribution transfer mechanism, color optimization with a straightforward MSE loss and sketch similarity optimization with a CLIP-based geometric similarity loss.
arXiv Detail & Related papers (2024-04-02T11:03:24Z) - Bridging 3D Gaussian and Mesh for Freeview Video Rendering [57.21847030980905]
GauMesh bridges the 3D Gaussian and Mesh for modeling and rendering the dynamic scenes.
We show that our approach adapts the appropriate type of primitives to represent the different parts of the dynamic scene.
arXiv Detail & Related papers (2024-03-18T04:01:26Z) - Learning Naturally Aggregated Appearance for Efficient 3D Editing [94.47518916521065]
We propose to replace the color field with an explicit 2D appearance aggregation, also called canonical image.
To avoid the distortion effect and facilitate convenient editing, we complement the canonical image with a projection field that maps 3D points onto 2D pixels for texture lookup.
Our representation, dubbed AGAP, well supports various ways of 3D editing (e.g., stylization, interactive drawing, and content extraction) with no need of re-optimization.
arXiv Detail & Related papers (2023-12-11T18:59:31Z) - 3D Surface Reconstruction in the Wild by Deforming Shape Priors from
Synthetic Data [24.97027425606138]
Reconstructing the underlying 3D surface of an object from a single image is a challenging problem.
We present a new method for joint category-specific 3D reconstruction and object pose estimation from a single image.
Our approach achieves state-of-the-art reconstruction performance across several real-world datasets.
arXiv Detail & Related papers (2023-02-24T20:37:27Z) - Differentiable Rendering for Pose Estimation in Proximity Operations [4.282159812965446]
Differentiable rendering aims to compute the derivative of the image rendering function with respect to the rendering parameters.
This paper presents a novel algorithm for 6-DoF pose estimation using a differentiable rendering pipeline.
arXiv Detail & Related papers (2022-12-24T06:12:16Z) - A Shading-Guided Generative Implicit Model for Shape-Accurate 3D-Aware
Image Synthesis [163.96778522283967]
We propose a shading-guided generative implicit model that is able to learn a starkly improved shape representation.
An accurate 3D shape should also yield a realistic rendering under different lighting conditions.
Our experiments on multiple datasets show that the proposed approach achieves photorealistic 3D-aware image synthesis.
arXiv Detail & Related papers (2021-10-29T10:53:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.