Improved Techniques for Training Single-Image GANs
- URL: http://arxiv.org/abs/2003.11512v2
- Date: Tue, 17 Nov 2020 10:55:13 GMT
- Title: Improved Techniques for Training Single-Image GANs
- Authors: Tobias Hinz, Matthew Fisher, Oliver Wang, Stefan Wermter
- Abstract summary: generative models can be learned from a single image, as opposed to from a large dataset.
We propose some best practices to train a model capable of generating realistic images from only a single sample.
Our model is up to six times faster to train, has fewer parameters, and can better capture the global structure of images.
- Score: 44.251222212306764
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently there has been an interest in the potential of learning generative
models from a single image, as opposed to from a large dataset. This task is of
practical significance, as it means that generative models can be used in
domains where collecting a large dataset is not feasible. However, training a
model capable of generating realistic images from only a single sample is a
difficult problem. In this work, we conduct a number of experiments to
understand the challenges of training these methods and propose some best
practices that we found allowed us to generate improved results over previous
work in this space. One key piece is that unlike prior single image generation
methods, we concurrently train several stages in a sequential multi-stage
manner, allowing us to learn models with fewer stages of increasing image
resolution. Compared to a recent state of the art baseline, our model is up to
six times faster to train, has fewer parameters, and can better capture the
global structure of images.
Related papers
- A Simple Approach to Unifying Diffusion-based Conditional Generation [63.389616350290595]
We introduce a simple, unified framework to handle diverse conditional generation tasks.
Our approach enables versatile capabilities via different inference-time sampling schemes.
Our model supports additional capabilities like non-spatially aligned and coarse conditioning.
arXiv Detail & Related papers (2024-10-15T09:41:43Z) - EfficientTrain++: Generalized Curriculum Learning for Efficient Visual Backbone Training [79.96741042766524]
We reformulate the training curriculum as a soft-selection function.
We show that exposing the contents of natural images can be readily achieved by the intensity of data augmentation.
The resulting method, EfficientTrain++, is simple, general, yet surprisingly effective.
arXiv Detail & Related papers (2024-05-14T17:00:43Z) - One-Shot Image Restoration [0.0]
Experimental results demonstrate the applicability, robustness and computational efficiency of the proposed approach for supervised image deblurring and super-resolution.
Our results showcase significant improvement of learning models' sample efficiency, generalization and time complexity.
arXiv Detail & Related papers (2024-04-26T14:03:23Z) - Is This Loss Informative? Faster Text-to-Image Customization by Tracking
Objective Dynamics [31.15864240403093]
We study the training dynamics of popular text-to-image personalization methods, aiming to speed them up.
We propose a simple drop-in early stopping criterion that only requires computing the regular training objective on a fixed set of inputs.
Our experiments on Stable Diffusion for 48 different concepts and three personalization methods demonstrate the competitive performance of our approach.
arXiv Detail & Related papers (2023-02-09T18:49:13Z) - KNN-Diffusion: Image Generation via Large-Scale Retrieval [40.6656651653888]
Learning to adapt enables several new capabilities.
Fine-tuning trained models to new samples can be achieved by simply adding them to the table.
Our diffusion-based model trains on images only, by leveraging a joint Text-Image multi-modal metric.
arXiv Detail & Related papers (2022-04-06T14:13:35Z) - InvGAN: Invertible GANs [88.58338626299837]
InvGAN, short for Invertible GAN, successfully embeds real images to the latent space of a high quality generative model.
This allows us to perform image inpainting, merging, and online data augmentation.
arXiv Detail & Related papers (2021-12-08T21:39:00Z) - Meta Internal Learning [88.68276505511922]
Internal learning for single-image generation is a framework, where a generator is trained to produce novel images based on a single image.
We propose a meta-learning approach that enables training over a collection of images, in order to model the internal statistics of the sample image more effectively.
Our results show that the models obtained are as suitable as single-image GANs for many common image applications.
arXiv Detail & Related papers (2021-10-06T16:27:38Z) - Pre-Trained Image Processing Transformer [95.93031793337613]
We develop a new pre-trained model, namely, image processing transformer (IPT)
We present to utilize the well-known ImageNet benchmark for generating a large amount of corrupted image pairs.
IPT model is trained on these images with multi-heads and multi-tails.
arXiv Detail & Related papers (2020-12-01T09:42:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.