PREIM3D: 3D Consistent Precise Image Attribute Editing from a Single
Image
- URL: http://arxiv.org/abs/2304.10263v1
- Date: Thu, 20 Apr 2023 12:33:56 GMT
- Title: PREIM3D: 3D Consistent Precise Image Attribute Editing from a Single
Image
- Authors: Jianhui Li, Jianmin Li, Haoji Zhang, Shilong Liu, Zhengyi Wang, Zihao
Xiao, Kaiwen Zheng, Jun Zhu
- Abstract summary: We study the 3D-aware image attribute editing problem in this paper.
Recent methods solved the problem by training a shared encoder to map images into a 3D generator's latent space.
We propose two novel methods, an alternating training scheme and a multi-view identity loss, to maintain 3D consistency.
- Score: 23.06474962139909
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the 3D-aware image attribute editing problem in this paper, which
has wide applications in practice. Recent methods solved the problem by
training a shared encoder to map images into a 3D generator's latent space or
by per-image latent code optimization and then edited images in the latent
space. Despite their promising results near the input view, they still suffer
from the 3D inconsistency of produced images at large camera poses and
imprecise image attribute editing, like affecting unspecified attributes during
editing. For more efficient image inversion, we train a shared encoder for all
images. To alleviate 3D inconsistency at large camera poses, we propose two
novel methods, an alternating training scheme and a multi-view identity loss,
to maintain 3D consistency and subject identity. As for imprecise image
editing, we attribute the problem to the gap between the latent space of real
images and that of generated images. We compare the latent space and inversion
manifold of GAN models and demonstrate that editing in the inversion manifold
can achieve better results in both quantitative and qualitative evaluations.
Extensive experiments show that our method produces more 3D consistent images
and achieves more precise image editing than previous work. Source code and
pretrained models can be found on our project page:
https://mybabyyh.github.io/Preim3D/
Related papers
- DragGaussian: Enabling Drag-style Manipulation on 3D Gaussian Representation [57.406031264184584]
DragGaussian is a 3D object drag-editing framework based on 3D Gaussian Splatting.
Our contributions include the introduction of a new task, the development of DragGaussian for interactive point-based 3D editing, and comprehensive validation of its effectiveness through qualitative and quantitative experiments.
arXiv Detail & Related papers (2024-05-09T14:34:05Z) - DGE: Direct Gaussian 3D Editing by Consistent Multi-view Editing [72.54566271694654]
We consider the problem of editing 3D objects and scenes based on open-ended language instructions.
A common approach to this problem is to use a 2D image generator or editor to guide the 3D editing process.
This process is often inefficient due to the need for iterative updates of costly 3D representations.
arXiv Detail & Related papers (2024-04-29T17:59:30Z) - View-Consistent 3D Editing with Gaussian Splatting [50.6460814430094]
View-consistent Editing (VcEdit) is a novel framework that seamlessly incorporates 3DGS into image editing processes.
By incorporating consistency modules into an iterative pattern, VcEdit proficiently resolves the issue of multi-view inconsistency.
arXiv Detail & Related papers (2024-03-18T15:22:09Z) - GaussCtrl: Multi-View Consistent Text-Driven 3D Gaussian Splatting Editing [38.948892064761914]
GaussCtrl is a text-driven method to edit a 3D scene reconstructed by the 3D Gaussian Splatting (3DGS)
Our key contribution is multi-view consistent editing, which enables editing all images together instead of iteratively editing one image.
arXiv Detail & Related papers (2024-03-13T17:35:28Z) - Learning Naturally Aggregated Appearance for Efficient 3D Editing [94.47518916521065]
We propose to replace the color field with an explicit 2D appearance aggregation, also called canonical image.
To avoid the distortion effect and facilitate convenient editing, we complement the canonical image with a projection field that maps 3D points onto 2D pixels for texture lookup.
Our representation, dubbed AGAP, well supports various ways of 3D editing (e.g., stylization, interactive drawing, and content extraction) with no need of re-optimization.
arXiv Detail & Related papers (2023-12-11T18:59:31Z) - InstructPix2NeRF: Instructed 3D Portrait Editing from a Single Image [25.076270175205593]
InstructPix2NeRF enables instructed 3D-aware portrait editing from a single open-world image with human instructions.
At its core lies a conditional latent 3D diffusion process that lifts 2D editing to 3D space by learning the correlation between the paired images' difference and the instructions via triplet data.
arXiv Detail & Related papers (2023-11-06T02:21:11Z) - Self-Supervised Geometry-Aware Encoder for Style-Based 3D GAN Inversion [115.82306502822412]
StyleGAN has achieved great progress in 2D face reconstruction and semantic editing via image inversion and latent editing.
A corresponding generic 3D GAN inversion framework is still missing, limiting the applications of 3D face reconstruction and semantic editing.
We study the challenging problem of 3D GAN inversion where a latent code is predicted given a single face image to faithfully recover its 3D shapes and detailed textures.
arXiv Detail & Related papers (2022-12-14T18:49:50Z) - EpiGRAF: Rethinking training of 3D GANs [60.38818140637367]
We show that it is possible to obtain a high-resolution 3D generator with SotA image quality by following a completely different route of simply training the model patch-wise.
The resulting model, named EpiGRAF, is an efficient, high-resolution, pure 3D generator.
arXiv Detail & Related papers (2022-06-21T17:08:23Z) - Differentiable Rendering: A Survey [22.35293459579154]
Differentiable rendering is a novel field which allows the gradients of 3D objects to be calculated and propagated through images.
This paper reviews existing literature and discusses the current state of differentiable rendering, its applications and open research problems.
arXiv Detail & Related papers (2020-06-22T08:14:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.