Learning 3D-aware Image Synthesis with Unknown Pose Distribution
- URL: http://arxiv.org/abs/2301.07702v2
- Date: Thu, 23 Mar 2023 12:25:12 GMT
- Title: Learning 3D-aware Image Synthesis with Unknown Pose Distribution
- Authors: Zifan Shi, Yujun Shen, Yinghao Xu, Sida Peng, Yiyi Liao, Sheng Guo,
Qifeng Chen, Dit-Yan Yeung
- Abstract summary: Existing methods for 3D-aware image synthesis largely depend on the 3D pose distribution pre-estimated on the training set.
This work proposes PoF3D that frees generative radiance fields from the requirements of 3D pose priors.
- Score: 68.62476998646866
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing methods for 3D-aware image synthesis largely depend on the 3D pose
distribution pre-estimated on the training set. An inaccurate estimation may
mislead the model into learning faulty geometry. This work proposes PoF3D that
frees generative radiance fields from the requirements of 3D pose priors. We
first equip the generator with an efficient pose learner, which is able to
infer a pose from a latent code, to approximate the underlying true pose
distribution automatically. We then assign the discriminator a task to learn
pose distribution under the supervision of the generator and to differentiate
real and synthesized images with the predicted pose as the condition. The
pose-free generator and the pose-aware discriminator are jointly trained in an
adversarial manner. Extensive results on a couple of datasets confirm that the
performance of our approach, regarding both image quality and geometry quality,
is on par with state of the art. To our best knowledge, PoF3D demonstrates the
feasibility of learning high-quality 3D-aware image synthesis without using 3D
pose priors for the first time.
Related papers
- MPL: Lifting 3D Human Pose from Multi-view 2D Poses [75.26416079541723]
We propose combining 2D pose estimation, for which large and rich training datasets exist, and 2D-to-3D pose lifting, using a transformer-based network.
Our experiments demonstrate decreases up to 45% in MPJPE errors compared to the 3D pose obtained by triangulating the 2D poses.
arXiv Detail & Related papers (2024-08-20T12:55:14Z) - Learning 3D-Aware GANs from Unposed Images with Template Feature Field [33.32761749864555]
This work targets learning 3D-aware GANs from unposed images.
We propose to perform on-the-fly pose estimation of training images with a learned template feature field (TeFF)
arXiv Detail & Related papers (2024-04-08T17:42:08Z) - SideGAN: 3D-Aware Generative Model for Improved Side-View Image
Synthesis [44.05449141767394]
We propose SideGAN, a novel 3D GAN training method to generate photo-realistic images irrespective of the camera pose.
We show that our approach enables 3D GANs to generate high-quality geometries and photo-realistic images irrespective of the camera pose.
arXiv Detail & Related papers (2023-09-19T07:38:05Z) - PoseTriplet: Co-evolving 3D Human Pose Estimation, Imitation, and
Hallucination under Self-supervision [102.48681650013698]
Existing self-supervised 3D human pose estimation schemes have largely relied on weak supervisions to guide the learning.
We propose a novel self-supervised approach that allows us to explicitly generate 2D-3D pose pairs for augmenting supervision.
This is made possible via introducing a reinforcement-learning-based imitator, which is learned jointly with a pose estimator alongside a pose hallucinator.
arXiv Detail & Related papers (2022-03-29T14:45:53Z) - CAMPARI: Camera-Aware Decomposed Generative Neural Radiance Fields [67.76151996543588]
We learn a 3D- and camera-aware generative model which faithfully recovers not only the image but also the camera data distribution.
At test time, our model generates images with explicit control over the camera as well as the shape and appearance of the scene.
arXiv Detail & Related papers (2021-03-31T17:59:24Z) - Unsupervised Cross-Modal Alignment for Multi-Person 3D Pose Estimation [52.94078950641959]
We present a deployment friendly, fast bottom-up framework for multi-person 3D human pose estimation.
We adopt a novel neural representation of multi-person 3D pose which unifies the position of person instances with their corresponding 3D pose representation.
We propose a practical deployment paradigm where paired 2D or 3D pose annotations are unavailable.
arXiv Detail & Related papers (2020-08-04T07:54:25Z) - Self-Supervised 3D Human Pose Estimation via Part Guided Novel Image
Synthesis [72.34794624243281]
We propose a self-supervised learning framework to disentangle variations from unlabeled video frames.
Our differentiable formalization, bridging the representation gap between the 3D pose and spatial part maps, allows us to operate on videos with diverse camera movements.
arXiv Detail & Related papers (2020-04-09T07:55:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.