The Euclidean Space is Evil: Hyperbolic Attribute Editing for Few-shot
Image Generation
- URL: http://arxiv.org/abs/2211.12347v2
- Date: Sat, 19 Aug 2023 11:20:04 GMT
- Title: The Euclidean Space is Evil: Hyperbolic Attribute Editing for Few-shot
Image Generation
- Authors: Lingxiao Li, Yi Zhang, Shuhui Wang
- Abstract summary: We propose Hyperbolic Attribute Editing(HAE) to generate diverse new images for an unseen category with only a few images.
Unlike other methods that work in Euclidean space, HAE captures the hierarchy among images using data from seen categories in hyperbolic space.
experiments and visualizations demonstrate that HAE is capable of not only generating images with promising quality and diversity using limited data but achieving a highly controllable and interpretable editing process.
- Score: 39.26386610133435
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few-shot image generation is a challenging task since it aims to generate
diverse new images for an unseen category with only a few images. Existing
methods suffer from the trade-off between the quality and diversity of
generated images. To tackle this problem, we propose Hyperbolic Attribute
Editing~(HAE), a simple yet effective method. Unlike other methods that work in
Euclidean space, HAE captures the hierarchy among images using data from seen
categories in hyperbolic space. Given a well-trained HAE, images of unseen
categories can be generated by moving the latent code of a given image toward
any meaningful directions in the Poincar\'e disk with a fixing radius. Most
importantly, the hyperbolic space allows us to control the semantic diversity
of the generated images by setting different radii in the disk. Extensive
experiments and visualizations demonstrate that HAE is capable of not only
generating images with promising quality and diversity using limited data but
achieving a highly controllable and interpretable editing process.
Related papers
- Diffusion Autoencoders for Few-shot Image Generation in Hyperbolic Space [35.516111930052105]
Few-shot image generation aims to generate diverse and high-quality images for an unseen class given only a few examples in that class.
We propose Hyperbolic Diffusion Autoencoders (HypDAE), a novel approach that operates in hyperbolic space to capture hierarchical relationships among images and texts from seen categories.
arXiv Detail & Related papers (2024-11-27T00:45:51Z) - The Stable Artist: Steering Semantics in Diffusion Latent Space [17.119616029527744]
We present the Stable Artist, an image editing approach enabling fine-grained control of the image generation process.
The main component is semantic guidance (SEGA) which steers the diffusion process along variable numbers of semantic directions.
SEGA enables probing of latent spaces to gain insights into the representation of concepts learned by the model.
arXiv Detail & Related papers (2022-12-12T16:21:24Z) - Reinforcing Generated Images via Meta-learning for One-Shot Fine-Grained
Visual Recognition [36.02360322125622]
We propose a meta-learning framework to combine generated images with original images, so that the resulting "hybrid" training images improve one-shot learning.
Our experiments demonstrate consistent improvement over baselines on one-shot fine-grained image classification benchmarks.
arXiv Detail & Related papers (2022-04-22T13:11:05Z) - Spatially Multi-conditional Image Generation [80.04130168156792]
We propose a novel neural architecture to address the problem of multi-conditional image generation.
The proposed method uses a transformer-like architecture operating pixel-wise, which receives the available labels as input tokens.
Our experiments on three benchmark datasets demonstrate the clear superiority of our method over the state-of-the-art and the compared baselines.
arXiv Detail & Related papers (2022-03-25T17:57:13Z) - Attribute Group Editing for Reliable Few-shot Image Generation [85.52840521454411]
We propose a new editing-based method, i.e., Attribute Group Editing (AGE), for few-shot image generation.
AGE examines the internal representation learned in GANs and identifies semantically meaningful directions.
arXiv Detail & Related papers (2022-03-16T06:54:09Z) - SMILE: Semantically-guided Multi-attribute Image and Layout Editing [154.69452301122175]
Attribute image manipulation has been a very active topic since the introduction of Generative Adversarial Networks (GANs)
We present a multimodal representation that handles all attributes, be it guided by random noise or images, while only using the underlying domain information of the target domain.
Our method is capable of adding, removing or changing either fine-grained or coarse attributes by using an image as a reference or by exploring the style distribution space.
arXiv Detail & Related papers (2020-10-05T20:15:21Z) - DVG-Face: Dual Variational Generation for Heterogeneous Face Recognition [85.94331736287765]
We formulate HFR as a dual generation problem, and tackle it via a novel Dual Variational Generation (DVG-Face) framework.
We integrate abundant identity information of large-scale visible data into the joint distribution.
Massive new diverse paired heterogeneous images with the same identity can be generated from noises.
arXiv Detail & Related papers (2020-09-20T09:48:24Z) - Exploiting Deep Generative Prior for Versatile Image Restoration and
Manipulation [181.08127307338654]
This work presents an effective way to exploit the image prior captured by a generative adversarial network (GAN) trained on large-scale natural images.
The deep generative prior (DGP) provides compelling results to restore missing semantics, e.g., color, patch, resolution, of various degraded images.
arXiv Detail & Related papers (2020-03-30T17:45:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.