Do 2D GANs Know 3D Shape? Unsupervised 3D shape reconstruction from 2D
Image GANs
- URL: http://arxiv.org/abs/2011.00844v4
- Date: Fri, 12 Mar 2021 04:04:05 GMT
- Title: Do 2D GANs Know 3D Shape? Unsupervised 3D shape reconstruction from 2D
Image GANs
- Authors: Xingang Pan, Bo Dai, Ziwei Liu, Chen Change Loy, Ping Luo
- Abstract summary: State-of-the-art 2D generative models like GANs show unprecedented quality in modeling the natural image manifold.
We present the first attempt to directly mine 3D geometric cues from an off-the-shelf 2D GAN that is trained on RGB images only.
- Score: 156.1209884183522
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Natural images are projections of 3D objects on a 2D image plane. While
state-of-the-art 2D generative models like GANs show unprecedented quality in
modeling the natural image manifold, it is unclear whether they implicitly
capture the underlying 3D object structures. And if so, how could we exploit
such knowledge to recover the 3D shapes of objects in the images? To answer
these questions, in this work, we present the first attempt to directly mine 3D
geometric cues from an off-the-shelf 2D GAN that is trained on RGB images only.
Through our investigation, we found that such a pre-trained GAN indeed contains
rich 3D knowledge and thus can be used to recover 3D shape from a single 2D
image in an unsupervised manner. The core of our framework is an iterative
strategy that explores and exploits diverse viewpoint and lighting variations
in the GAN image manifold. The framework does not require 2D keypoint or 3D
annotations, or strong assumptions on object shapes (e.g. shapes are
symmetric), yet it successfully recovers 3D shapes with high precision for
human faces, cats, cars, and buildings. The recovered 3D shapes immediately
allow high-quality image editing like relighting and object rotation. We
quantitatively demonstrate the effectiveness of our approach compared to
previous methods in both 3D shape reconstruction and face rotation. Our code is
available at https://github.com/XingangPan/GAN2Shape.
Related papers
- AG3D: Learning to Generate 3D Avatars from 2D Image Collections [96.28021214088746]
We propose a new adversarial generative model of realistic 3D people from 2D images.
Our method captures shape and deformation of the body and loose clothing by adopting a holistic 3D generator.
We experimentally find that our method outperforms previous 3D- and articulation-aware methods in terms of geometry and appearance.
arXiv Detail & Related papers (2023-05-03T17:56:24Z) - Self-Supervised Geometry-Aware Encoder for Style-Based 3D GAN Inversion [115.82306502822412]
StyleGAN has achieved great progress in 2D face reconstruction and semantic editing via image inversion and latent editing.
A corresponding generic 3D GAN inversion framework is still missing, limiting the applications of 3D face reconstruction and semantic editing.
We study the challenging problem of 3D GAN inversion where a latent code is predicted given a single face image to faithfully recover its 3D shapes and detailed textures.
arXiv Detail & Related papers (2022-12-14T18:49:50Z) - XDGAN: Multi-Modal 3D Shape Generation in 2D Space [60.46777591995821]
We propose a novel method to convert 3D shapes into compact 1-channel geometry images and leverage StyleGAN3 and image-to-image translation networks to generate 3D objects in 2D space.
The generated geometry images are quick to convert to 3D meshes, enabling real-time 3D object synthesis, visualization and interactive editing.
We show both quantitatively and qualitatively that our method is highly effective at various tasks such as 3D shape generation, single view reconstruction and shape manipulation, while being significantly faster and more flexible compared to recent 3D generative models.
arXiv Detail & Related papers (2022-10-06T15:54:01Z) - GAN2X: Non-Lambertian Inverse Rendering of Image GANs [85.76426471872855]
We present GAN2X, a new method for unsupervised inverse rendering that only uses unpaired images for training.
Unlike previous Shape-from-GAN approaches that mainly focus on 3D shapes, we take the first attempt to also recover non-Lambertian material properties by exploiting the pseudo paired data generated by a GAN.
Experiments demonstrate that GAN2X can accurately decompose 2D images to 3D shape, albedo, and specular properties for different object categories, and achieves the state-of-the-art performance for unsupervised single-view 3D face reconstruction.
arXiv Detail & Related papers (2022-06-18T16:58:49Z) - Fully Understanding Generic Objects: Modeling, Segmentation, and
Reconstruction [33.95791350070165]
Inferring 3D structure of a generic object from a 2D image is a long-standing objective of computer vision.
We take an alternative approach with semi-supervised learning. That is, for a 2D image of a generic object, we decompose it into latent representations of category, shape and albedo.
We show that the complete shape and albedo modeling enables us to leverage real 2D images in both modeling and model fitting.
arXiv Detail & Related papers (2021-04-02T02:39:29Z) - Lifting 2D StyleGAN for 3D-Aware Face Generation [52.8152883980813]
We propose a framework, called LiftedGAN, that disentangles and lifts a pre-trained StyleGAN2 for 3D-aware face generation.
Our model is "3D-aware" in the sense that it is able to (1) disentangle the latent space of StyleGAN2 into texture, shape, viewpoint, lighting and (2) generate 3D components for synthetic images.
arXiv Detail & Related papers (2020-11-26T05:02:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.