BachGAN: High-Resolution Image Synthesis from Salient Object Layout
- URL: http://arxiv.org/abs/2003.11690v2
- Date: Fri, 27 Mar 2020 20:53:24 GMT
- Title: BachGAN: High-Resolution Image Synthesis from Salient Object Layout
- Authors: Yandong Li, Yu Cheng, Zhe Gan, Licheng Yu, Liqiang Wang, and Jingjing
Liu
- Abstract summary: We propose a new task towards more practical application for image generation - high-quality image synthesis from salient object layout.
Two main challenges spring from this new task: (i) how to generate fine-grained details and realistic textures without segmentation map input; and (ii) how to create a background and weave it seamlessly into standalone objects.
By generating the hallucinated background representation dynamically, our model can synthesize high-resolution images with both photo-realistic foreground and integral background.
- Score: 78.51640906030244
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a new task towards more practical application for image generation
- high-quality image synthesis from salient object layout. This new setting
allows users to provide the layout of salient objects only (i.e., foreground
bounding boxes and categories), and lets the model complete the drawing with an
invented background and a matching foreground. Two main challenges spring from
this new task: (i) how to generate fine-grained details and realistic textures
without segmentation map input; and (ii) how to create a background and weave
it seamlessly into standalone objects. To tackle this, we propose Background
Hallucination Generative Adversarial Network (BachGAN), which first selects a
set of segmentation maps from a large candidate pool via a background retrieval
module, then encodes these candidate layouts via a background fusion module to
hallucinate a suitable background for the given objects. By generating the
hallucinated background representation dynamically, our model can synthesize
high-resolution images with both photo-realistic foreground and integral
background. Experiments on Cityscapes and ADE20K datasets demonstrate the
advantage of BachGAN over existing methods, measured on both visual fidelity of
generated images and visual alignment between output images and input layouts.
Related papers
- GroundingBooth: Grounding Text-to-Image Customization [17.185571339157075]
We introduce GroundingBooth, a framework that achieves zero-shot instance-level spatial grounding on both foreground subjects and background objects.
Our proposed text-image grounding module and masked cross-attention layer allow us to generate personalized images with both accurate layout alignment and identity preservation.
arXiv Detail & Related papers (2024-09-13T03:40:58Z) - Painterly Image Harmonization via Adversarial Residual Learning [37.78751164466694]
painterly image aims to transfer the style of background painting to the foreground object.
In this work, we employ adversarial learning to bridge the domain gap between foreground feature map and background feature map.
arXiv Detail & Related papers (2023-11-15T01:53:46Z) - BATINet: Background-Aware Text to Image Synthesis and Manipulation
Network [12.924990882126105]
We analyzed a novel Background-Aware Text2Image (BAT2I) task in which the generated content matches the input background.
We proposed a Background-Aware Text to Image synthesis and manipulation Network (BATINet), which contains two key components.
We demonstrated through qualitative and quantitative evaluations on the CUB dataset that the proposed model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2023-08-11T03:22:33Z) - SIEDOB: Semantic Image Editing by Disentangling Object and Background [5.149242555705579]
We propose a novel paradigm for semantic image editing.
textbfSIEDOB, the core idea of which is to explicitly leverage several heterogeneousworks for objects and backgrounds.
We conduct extensive experiments on Cityscapes and ADE20K-Room datasets and exhibit that our method remarkably outperforms the baselines.
arXiv Detail & Related papers (2023-03-23T06:17:23Z) - SceneComposer: Any-Level Semantic Image Synthesis [80.55876413285587]
We propose a new framework for conditional image synthesis from semantic layouts of any precision levels.
The framework naturally reduces to text-to-image (T2I) at the lowest level with no shape information, and it becomes segmentation-to-image (S2I) at the highest level.
We introduce several novel techniques to address the challenges coming with this new setup.
arXiv Detail & Related papers (2022-11-21T18:59:05Z) - Deep Rectangling for Image Stitching: A Learning Baseline [57.76737888499145]
We build the first image stitching rectangling dataset with a large diversity in irregular boundaries and scenes.
Experiments demonstrate our superiority over traditional methods both quantitatively and qualitatively.
arXiv Detail & Related papers (2022-03-08T03:34:10Z) - Content-aware Warping for View Synthesis [110.54435867693203]
We propose content-aware warping, which adaptively learns the weights for pixels of a relatively large neighborhood from their contextual information via a lightweight neural network.
Based on this learnable warping module, we propose a new end-to-end learning-based framework for novel view synthesis from two source views.
Experimental results on structured light field datasets with wide baselines and unstructured multi-view datasets show that the proposed method significantly outperforms state-of-the-art methods both quantitatively and visually.
arXiv Detail & Related papers (2022-01-22T11:35:05Z) - Learned Spatial Representations for Few-shot Talking-Head Synthesis [68.3787368024951]
We propose a novel approach for few-shot talking-head synthesis.
We show that this disentangled representation leads to a significant improvement over previous methods.
arXiv Detail & Related papers (2021-04-29T17:59:42Z) - Generating Object Stamps [47.20601520671103]
We present an algorithm to generate diverse foreground objects and composite them into background images using a GAN architecture.
Our results on the challenging COCO dataset show improved overall quality and diversity compared to state-of-the-art object insertion approaches.
arXiv Detail & Related papers (2020-01-01T14:36:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.