StyleRes: Transforming the Residuals for Real Image Editing with
StyleGAN
- URL: http://arxiv.org/abs/2212.14359v1
- Date: Thu, 29 Dec 2022 16:14:09 GMT
- Title: StyleRes: Transforming the Residuals for Real Image Editing with
StyleGAN
- Authors: Hamza Pehlivan, Yusuf Dalva, Aysegul Dundar
- Abstract summary: Inverting real images into StyleGAN's latent space is an extensively studied problem.
Trade-off between the image reconstruction fidelity and image editing quality remains an open challenge.
We present a novel image inversion framework and a training pipeline to achieve high-fidelity image inversion with high-quality editing.
- Score: 4.7590051176368915
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a novel image inversion framework and a training pipeline to
achieve high-fidelity image inversion with high-quality attribute editing.
Inverting real images into StyleGAN's latent space is an extensively studied
problem, yet the trade-off between the image reconstruction fidelity and image
editing quality remains an open challenge. The low-rate latent spaces are
limited in their expressiveness power for high-fidelity reconstruction. On the
other hand, high-rate latent spaces result in degradation in editing quality.
In this work, to achieve high-fidelity inversion, we learn residual features in
higher latent codes that lower latent codes were not able to encode. This
enables preserving image details in reconstruction. To achieve high-quality
editing, we learn how to transform the residual features for adapting to
manipulations in latent codes. We train the framework to extract residual
features and transform them via a novel architecture pipeline and cycle
consistency losses. We run extensive experiments and compare our method with
state-of-the-art inversion methods. Qualitative metrics and visual comparisons
show significant improvements. Code: https://github.com/hamzapehlivan/StyleRes
Related papers
- The Devil is in the Details: StyleFeatureEditor for Detail-Rich StyleGAN Inversion and High Quality Image Editing [3.58736715327935]
We introduce StyleFeatureEditor, a novel method that enables editing in both w-latents and F-latents.
We also present a new training pipeline specifically designed to train our model to accurately edit F-latents.
Our method is compared with state-of-the-art encoding approaches, demonstrating that our model excels in terms of reconstruction quality.
arXiv Detail & Related papers (2024-06-15T11:28:32Z) - ENTED: Enhanced Neural Texture Extraction and Distribution for
Reference-based Blind Face Restoration [51.205673783866146]
We present ENTED, a new framework for blind face restoration that aims to restore high-quality and realistic portrait images.
We utilize a texture extraction and distribution framework to transfer high-quality texture features between the degraded input and reference image.
The StyleGAN-like architecture in our framework requires high-quality latent codes to generate realistic images.
arXiv Detail & Related papers (2024-01-13T04:54:59Z) - Warping the Residuals for Image Editing with StyleGAN [5.733811543584874]
StyleGAN models show editing capabilities via their semantically interpretable latent organizations.
Many works have been proposed for inverting images into StyleGAN's latent space.
We present a novel image inversion architecture that extracts high-rate latent features and includes a flow estimation module.
arXiv Detail & Related papers (2023-12-18T18:24:18Z) - In-Domain GAN Inversion for Faithful Reconstruction and Editability [132.68255553099834]
We propose in-domain GAN inversion, which consists of a domain-guided domain-regularized and a encoder to regularize the inverted code in the native latent space of the pre-trained GAN model.
We make comprehensive analyses on the effects of the encoder structure, the starting inversion point, as well as the inversion parameter space, and observe the trade-off between the reconstruction quality and the editing property.
arXiv Detail & Related papers (2023-09-25T08:42:06Z) - Diverse Inpainting and Editing with GAN Inversion [4.234367850767171]
Recent inversion methods have shown that real images can be inverted into StyleGAN's latent space.
In this paper, we tackle an even more difficult task, inverting erased images into GAN's latent space for realistic inpaintings and editings.
arXiv Detail & Related papers (2023-07-27T17:41:36Z) - Towards Accurate Image Coding: Improved Autoregressive Image Generation
with Dynamic Vector Quantization [73.52943587514386]
Existing vector quantization (VQ) based autoregressive models follow a two-stage generation paradigm.
We propose a novel two-stage framework: (1) Dynamic-Quantization VAE (DQ-VAE) which encodes image regions into variable-length codes based their information densities for accurate representation.
arXiv Detail & Related papers (2023-05-19T14:56:05Z) - Overparameterization Improves StyleGAN Inversion [66.8300251627992]
Existing inversion approaches obtain promising yet imperfect results.
We show that this allows us to obtain near-perfect image reconstruction without the need for encoders.
Our approach also retains editability, which we demonstrate by realistically interpolating between images.
arXiv Detail & Related papers (2022-05-12T18:42:43Z) - High-Fidelity GAN Inversion for Image Attribute Editing [61.966946442222735]
We present a novel high-fidelity generative adversarial network (GAN) inversion framework that enables attribute editing with image-specific details well-preserved.
With a low bit-rate latent code, previous works have difficulties in preserving high-fidelity details in reconstructed and edited images.
We propose a distortion consultation approach that employs a distortion map as a reference for high-fidelity reconstruction.
arXiv Detail & Related papers (2021-09-14T11:23:48Z) - Improved StyleGAN Embedding: Where are the Good Latents? [43.780075713984935]
StyleGAN is able to produce photorealistic images that are almost indistinguishable from real ones.
The reverse problem of finding an embedding for a given image poses a challenge.
In this paper, we address the problem of finding an embedding that both reconstructs images and also supports image editing tasks.
arXiv Detail & Related papers (2020-12-13T18:01:24Z) - In-Domain GAN Inversion for Real Image Editing [56.924323432048304]
A common practice of feeding a real image to a trained GAN generator is to invert it back to a latent code.
Existing inversion methods typically focus on reconstructing the target image by pixel values yet fail to land the inverted code in the semantic domain of the original latent space.
We propose an in-domain GAN inversion approach, which faithfully reconstructs the input image and ensures the inverted code to be semantically meaningful for editing.
arXiv Detail & Related papers (2020-03-31T18:20:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.