Adversarial Image Composition with Auxiliary Illumination
- URL: http://arxiv.org/abs/2009.08255v2
- Date: Sat, 9 Jan 2021 15:05:42 GMT
- Title: Adversarial Image Composition with Auxiliary Illumination
- Authors: Fangneng Zhan, Shijian Lu, Changgong Zhang, Feiying Ma, Xuansong Xie
- Abstract summary: We propose an Adversarial Image Composition Net (AIC-Net) that achieves realistic image composition.
A novel branched generation mechanism is proposed, which disentangles the generation of shadows and the transfer of foreground styles.
Experiments on pedestrian and car composition tasks show that the proposed AIC-Net achieves superior composition performance.
- Score: 53.89445873577062
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dealing with the inconsistency between a foreground object and a background
image is a challenging task in high-fidelity image composition.
State-of-the-art methods strive to harmonize the composed image by adapting the
style of foreground objects to be compatible with the background image, whereas
the potential shadow of foreground objects within the composed image which is
critical to the composition realism is largely neglected. In this paper, we
propose an Adversarial Image Composition Net (AIC-Net) that achieves realistic
image composition by considering potential shadows that the foreground object
projects in the composed image. A novel branched generation mechanism is
proposed, which disentangles the generation of shadows and the transfer of
foreground styles for optimal accomplishment of the two tasks simultaneously. A
differentiable spatial transformation module is designed which bridges the
local harmonization and the global harmonization to achieve their joint
optimization effectively. Extensive experiments on pedestrian and car
composition tasks show that the proposed AIC-Net achieves superior composition
performance qualitatively and quantitatively.
Related papers
- FreeCompose: Generic Zero-Shot Image Composition with Diffusion Prior [50.0535198082903]
We offer a novel approach to image composition, which integrates multiple input images into a single, coherent image.
We showcase the potential of utilizing the powerful generative prior inherent in large-scale pre-trained diffusion models to accomplish generic image composition.
arXiv Detail & Related papers (2024-07-06T03:35:43Z) - Repositioning the Subject within Image [78.8467524191102]
We introduce an innovative dynamic manipulation task, subject repositioning.
This task involves relocating a user-specified subject to a desired position while preserving the image's fidelity.
Our research reveals that the fundamental sub-tasks of subject repositioning can be effectively reformulated as a unified, prompt-guided inpainting task.
arXiv Detail & Related papers (2024-01-30T10:04:49Z) - Semantic-aware Network for Aerial-to-Ground Image Synthesis [42.360670351361584]
We propose a novel framework to explore the challenges by imposing enhanced structural alignment and semantic awareness.
We introduce a novel semantic-attentive feature transformation module that allows to reconstruct the complex geographic structures.
We also propose semantic-aware loss functions by leveraging a pre-trained segmentation network.
arXiv Detail & Related papers (2023-08-14T05:37:07Z) - Cones 2: Customizable Image Synthesis with Multiple Subjects [50.54010141032032]
We study how to efficiently represent a particular subject as well as how to appropriately compose different subjects.
By rectifying the activations in the cross-attention map, the layout appoints and separates the location of different subjects in the image.
arXiv Detail & Related papers (2023-05-30T18:00:06Z) - ObjectStitch: Generative Object Compositing [43.206123360578665]
We propose a self-supervised framework for object compositing using conditional diffusion models.
Our framework can transform the viewpoint, geometry, color and shadow of the generated object while requiring no manual labeling.
Our method outperforms relevant baselines in both realism and faithfulness of the synthesized result images in a user study on various real-world images.
arXiv Detail & Related papers (2022-12-02T02:15:13Z) - A Generative Adversarial Framework for Optimizing Image Matting and
Harmonization Simultaneously [7.541357996797061]
We propose a new Generative Adversarial (GAN) framework which optimizing the matting network and the harmonization network based on a self-attention discriminator.
Our dataset and dataset generating pipeline can be found in urlhttps://git.io/HaMaGAN
arXiv Detail & Related papers (2021-08-13T06:48:14Z) - Deep Image Compositing [93.75358242750752]
We propose a new method which can automatically generate high-quality image composites without any user input.
Inspired by Laplacian pyramid blending, a dense-connected multi-stream fusion network is proposed to effectively fuse the information from the foreground and background images.
Experiments show that the proposed method can automatically generate high-quality composites and outperforms existing methods both qualitatively and quantitatively.
arXiv Detail & Related papers (2020-11-04T06:12:24Z) - Hierarchy Composition GAN for High-fidelity Image Synthesis [57.32311953820988]
This paper presents an innovative Hierarchical Composition GAN (HIC-GAN)
HIC-GAN incorporates image synthesis in geometry and appearance domains into an end-to-end trainable network.
Experiments on scene text image synthesis, portrait editing and indoor rendering tasks show that the proposed HIC-GAN achieves superior synthesis performance qualitatively and quantitatively.
arXiv Detail & Related papers (2019-05-12T11:11:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.