Local and Global GANs with Semantic-Aware Upsampling for Image
Generation
- URL: http://arxiv.org/abs/2203.00047v1
- Date: Mon, 28 Feb 2022 19:24:25 GMT
- Title: Local and Global GANs with Semantic-Aware Upsampling for Image
Generation
- Authors: Hao Tang, Ling Shao, Philip H.S. Torr, Nicu Sebe
- Abstract summary: We consider generating images using local context.
We propose a class-specific generative network using semantic maps as guidance.
Lastly, we propose a novel semantic-aware upsampling method.
- Score: 201.39323496042527
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we address the task of semantic-guided image generation. One
challenge common to most existing image-level generation methods is the
difficulty in generating small objects and detailed local textures. To address
this, in this work we consider generating images using local context. As such,
we design a local class-specific generative network using semantic maps as
guidance, which separately constructs and learns subgenerators for different
classes, enabling it to capture finer details. To learn more discriminative
class-specific feature representations for the local generation, we also
propose a novel classification module. To combine the advantages of both global
image-level and local class-specific generation, a joint generation network is
designed with an attention fusion module and a dual-discriminator structure
embedded. Lastly, we propose a novel semantic-aware upsampling method, which
has a larger receptive field and can take far-away pixels that are semantically
related for feature upsampling, enabling it to better preserve semantic
consistency for instances with the same semantic labels. Extensive experiments
on two image generation tasks show the superior performance of the proposed
method. State-of-the-art results are established by large margins on both tasks
and on nine challenging public benchmarks. The source code and trained models
are available at https://github.com/Ha0Tang/LGGAN.
Related papers
- Dual Pyramid Generative Adversarial Networks for Semantic Image
Synthesis [94.76988562653845]
The goal of semantic image synthesis is to generate photo-realistic images from semantic label maps.
Current state-of-the-art approaches, however, still struggle to generate realistic objects in images at various scales.
We propose a Dual Pyramid Generative Adversarial Network (DP-GAN) that learns the conditioning of spatially-adaptive normalization blocks at all scales jointly.
arXiv Detail & Related papers (2022-10-08T18:45:44Z) - Boosting Image Outpainting with Semantic Layout Prediction [18.819765707811904]
We train a GAN to extend regions in semantic segmentation domain instead of image domain.
Another GAN model is trained to synthesize real images based on the extended semantic layouts.
Our approach can handle semantic clues more easily and hence works better in complex scenarios.
arXiv Detail & Related papers (2021-10-18T13:09:31Z) - Collaging Class-specific GANs for Semantic Image Synthesis [68.87294033259417]
We propose a new approach for high resolution semantic image synthesis.
It consists of one base image generator and multiple class-specific generators.
Experiments show that our approach can generate high quality images in high resolution.
arXiv Detail & Related papers (2021-10-08T17:46:56Z) - Improving Few-shot Learning with Weakly-supervised Object Localization [24.3569501375842]
We propose a novel framework that generates class representations by extracting features from class-relevant regions of the images.
Our method outperforms the baseline few-shot model in miniImageNet and tieredImageNet benchmarks.
arXiv Detail & Related papers (2021-05-25T07:39:32Z) - Semantic Segmentation with Generative Models: Semi-Supervised Learning
and Strong Out-of-Domain Generalization [112.68171734288237]
We propose a novel framework for discriminative pixel-level tasks using a generative model of both images and labels.
We learn a generative adversarial network that captures the joint image-label distribution and is trained efficiently using a large set of unlabeled images.
We demonstrate strong in-domain performance compared to several baselines, and are the first to showcase extreme out-of-domain generalization.
arXiv Detail & Related papers (2021-04-12T21:41:25Z) - Learning to Discover Multi-Class Attentional Regions for Multi-Label
Image Recognition [20.2935275611948]
We propose a simple but efficient two-stream framework to recognize multi-category objects from global image to local regions.
To bridge the gap between global and local streams, we propose a multi-class attentional region module.
Our method can efficiently and effectively recognize multi-class objects with an affordable computation cost and a parameter-free region localization module.
arXiv Detail & Related papers (2020-07-03T15:22:46Z) - OneGAN: Simultaneous Unsupervised Learning of Conditional Image
Generation, Foreground Segmentation, and Fine-Grained Clustering [100.32273175423146]
We present a method for simultaneously learning, in an unsupervised manner, a conditional image generator, foreground extraction and segmentation, and object removal and background completion.
The method combines a Geneversarative Adrial Network and a Variational Auto-Encoder, with multiple encoders, generators and discriminators, and benefits from solving all tasks at once.
arXiv Detail & Related papers (2019-12-31T18:15:58Z) - Local Class-Specific and Global Image-Level Generative Adversarial
Networks for Semantic-Guided Scene Generation [135.4660201856059]
We consider learning the scene generation in a local context, and design a local class-specific generative network with semantic maps as a guidance.
To learn more discrimi class-specific feature representations for the local generation, a novel classification module is also proposed.
Experiments on two scene image generation tasks show superior generation performance of the proposed model.
arXiv Detail & Related papers (2019-12-27T16:14:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.