BlendGAN: Learning and Blending the Internal Distributions of Single
Images by Spatial Image-Identity Conditioning
- URL: http://arxiv.org/abs/2212.01589v1
- Date: Sat, 3 Dec 2022 10:38:27 GMT
- Title: BlendGAN: Learning and Blending the Internal Distributions of Single
Images by Spatial Image-Identity Conditioning
- Authors: Idan Kligvasser, Tamar Rott Shaham, Noa Alkobi and Tomer Michaeli
- Abstract summary: Single image generative methods are designed to learn the internal patch distribution of a single natural image at multiple scales.
We introduce an extended framework, which allows to simultaneously learn the internal distributions of several images.
Our BlendGAN opens the door to applications that are not supported by single-image models.
- Score: 37.21764919074815
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Training a generative model on a single image has drawn significant attention
in recent years. Single image generative methods are designed to learn the
internal patch distribution of a single natural image at multiple scales. These
models can be used for drawing diverse samples that semantically resemble the
training image, as well as for solving many image editing and restoration tasks
that involve that particular image. Here, we introduce an extended framework,
which allows to simultaneously learn the internal distributions of several
images, by using a single model with spatially varying image-identity
conditioning. Our BlendGAN opens the door to applications that are not
supported by single-image models, including morphing, melding, and
structure-texture fusion between two or more arbitrary images.
Related papers
- A Simple Approach to Unifying Diffusion-based Conditional Generation [63.389616350290595]
We introduce a simple, unified framework to handle diverse conditional generation tasks.
Our approach enables versatile capabilities via different inference-time sampling schemes.
Our model supports additional capabilities like non-spatially aligned and coarse conditioning.
arXiv Detail & Related papers (2024-10-15T09:41:43Z) - Many-to-many Image Generation with Auto-regressive Diffusion Models [59.5041405824704]
This paper introduces a domain-general framework for many-to-many image generation, capable of producing interrelated image series from a given set of images.
We present MIS, a novel large-scale multi-image dataset, containing 12M synthetic multi-image samples, each with 25 interconnected images.
We learn M2M, an autoregressive model for many-to-many generation, where each image is modeled within a diffusion framework.
arXiv Detail & Related papers (2024-04-03T23:20:40Z) - A Method for Training-free Person Image Picture Generation [4.043367784553845]
A Character Image Feature model is proposed in this paper.
It enables the user to use the process by simply providing a picture of the character to make the image of the character in the generated image match the expectation.
The proposed model can be conveniently incorporated into the Stable Diffusion generation process without modifying the model's or used in combination with Stable Diffusion as a joint model.
arXiv Detail & Related papers (2023-05-16T21:46:28Z) - SinDiffusion: Learning a Diffusion Model from a Single Natural Image [159.4285444680301]
We present SinDiffusion, leveraging denoising diffusion models to capture internal distribution of patches from a single natural image.
It is based on two core designs. First, SinDiffusion is trained with a single model at a single scale instead of multiple models with progressive growing of scales.
Second, we identify that a patch-level receptive field of the diffusion network is crucial and effective for capturing the image's patch statistics.
arXiv Detail & Related papers (2022-11-22T18:00:03Z) - FewGAN: Generating from the Joint Distribution of a Few Images [95.6635227371479]
We introduce FewGAN, a generative model for generating novel, high-quality and diverse images.
FewGAN is a hierarchical patch-GAN that applies quantization at the first coarse scale, followed by a pyramid of residual fully convolutional GANs at finer scales.
In an extensive set of experiments, it is shown that FewGAN outperforms baselines both quantitatively and qualitatively.
arXiv Detail & Related papers (2022-07-18T07:11:28Z) - EdiBERT, a generative model for image editing [12.605607949417033]
EdiBERT is a bi-directional transformer trained in the discrete latent space built by a vector-quantized auto-encoder.
We show that the resulting model matches state-of-the-art performances on a wide variety of tasks.
arXiv Detail & Related papers (2021-11-30T10:23:06Z) - Meta Internal Learning [88.68276505511922]
Internal learning for single-image generation is a framework, where a generator is trained to produce novel images based on a single image.
We propose a meta-learning approach that enables training over a collection of images, in order to model the internal statistics of the sample image more effectively.
Our results show that the models obtained are as suitable as single-image GANs for many common image applications.
arXiv Detail & Related papers (2021-10-06T16:27:38Z) - Enhance Images as You Like with Unpaired Learning [8.104571453311442]
We propose a lightweight one-path conditional generative adversarial network (cGAN) to learn a one-to-many relation from low-light to normal-light image space.
Our network learns to generate a collection of enhanced images from a given input conditioned on various reference images.
Our model achieves competitive visual and quantitative results on par with fully supervised methods on both noisy and clean datasets.
arXiv Detail & Related papers (2021-10-04T03:00:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.