SCS-Co: Self-Consistent Style Contrastive Learning for Image
Harmonization
- URL: http://arxiv.org/abs/2204.13962v1
- Date: Fri, 29 Apr 2022 09:22:01 GMT
- Title: SCS-Co: Self-Consistent Style Contrastive Learning for Image
Harmonization
- Authors: Yucheng Hang, Bin Xia, Wenming Yang, Qingmin Liao
- Abstract summary: We propose a self-consistent style contrastive learning scheme (SCS-Co) for image harmonization.
By dynamically generating multiple negative samples, our SCS-Co can learn more distortion knowledge and well regularize the generated harmonized image.
In addition, we propose a background-attentional adaptive instance normalization (BAIN) to achieve an attention-weighted background feature distribution.
- Score: 29.600429707123645
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Image harmonization aims to achieve visual consistency in composite images by
adapting a foreground to make it compatible with a background. However,
existing methods always only use the real image as the positive sample to guide
the training, and at most introduce the corresponding composite image as a
single negative sample for an auxiliary constraint, which leads to limited
distortion knowledge, and further causes a too large solution space, making the
generated harmonized image distorted. Besides, none of them jointly constrain
from the foreground self-style and foreground-background style consistency,
which exacerbates this problem. Moreover, recent region-aware adaptive instance
normalization achieves great success but only considers the global background
feature distribution, making the aligned foreground feature distribution
biased. To address these issues, we propose a self-consistent style contrastive
learning scheme (SCS-Co). By dynamically generating multiple negative samples,
our SCS-Co can learn more distortion knowledge and well regularize the
generated harmonized image in the style representation space from two aspects
of the foreground self-style and foreground-background style consistency,
leading to a more photorealistic visual result. In addition, we propose a
background-attentional adaptive instance normalization (BAIN) to achieve an
attention-weighted background feature distribution according to the
foreground-background feature similarity. Experiments demonstrate the
superiority of our method over other state-of-the-art methods in both
quantitative comparison and visual analysis.
Related papers
- ZePo: Zero-Shot Portrait Stylization with Faster Sampling [61.14140480095604]
This paper presents an inversion-free portrait stylization framework based on diffusion models that accomplishes content and style feature fusion in merely four sampling steps.
We propose a feature merging strategy to amalgamate redundant features in Consistency Features, thereby reducing the computational load of attention control.
arXiv Detail & Related papers (2024-08-10T08:53:41Z) - FreeCompose: Generic Zero-Shot Image Composition with Diffusion Prior [50.0535198082903]
We offer a novel approach to image composition, which integrates multiple input images into a single, coherent image.
We showcase the potential of utilizing the powerful generative prior inherent in large-scale pre-trained diffusion models to accomplish generic image composition.
arXiv Detail & Related papers (2024-07-06T03:35:43Z) - Fine-grained Image-to-LiDAR Contrastive Distillation with Visual Foundation Models [55.99654128127689]
Visual Foundation Models (VFMs) are used to enhance 3D representation learning.
VFMs generate semantic labels for weakly-supervised pixel-to-point contrastive distillation.
We adapt sampling probabilities of points to address imbalances in spatial distribution and category frequency.
arXiv Detail & Related papers (2024-05-23T07:48:19Z) - Intrinsic Harmonization for Illumination-Aware Compositing [0.7366405857677227]
We introduce a self-supervised illumination harmonization approach formulated in the intrinsic image domain.
First, we estimate a simple global lighting model from mid-level vision representations to generate a rough shading for the foreground region.
A network then refines this inferred shading to generate a re-shading that aligns with the background scene.
arXiv Detail & Related papers (2023-12-06T18:59:03Z) - FreePIH: Training-Free Painterly Image Harmonization with Diffusion
Model [19.170302996189335]
Our FreePIH method tames the denoising process as a plug-in module for foreground image style transfer.
We make use of multi-scale features to enforce the consistency of the content and stability of the foreground objects in the latent space.
Our method can surpass representative baselines by large margins.
arXiv Detail & Related papers (2023-11-25T04:23:49Z) - Parallax-Tolerant Unsupervised Deep Image Stitching [57.76737888499145]
We propose UDIS++, a parallax-tolerant unsupervised deep image stitching technique.
First, we propose a robust and flexible warp to model the image registration from global homography to local thin-plate spline motion.
To further eliminate the parallax artifacts, we propose to composite the stitched image seamlessly by unsupervised learning for seam-driven composition masks.
arXiv Detail & Related papers (2023-02-16T10:40:55Z) - Image Harmonization with Region-wise Contrastive Learning [51.309905690367835]
We propose a novel image harmonization framework with external style fusion and region-wise contrastive learning scheme.
Our method attempts to bring together corresponding positive and negative samples by maximizing the mutual information between the foreground and background styles.
arXiv Detail & Related papers (2022-05-27T15:46:55Z) - SSH: A Self-Supervised Framework for Image Harmonization [97.16345684998788]
We propose a novel Self-Supervised Harmonization framework (SSH) that can be trained using just "free" natural images without being edited.
Our results show that the proposedSSH outperforms previous state-of-the-art methods in terms of reference metrics, visual quality, and subject user study.
arXiv Detail & Related papers (2021-08-15T19:51:33Z) - Region-aware Adaptive Instance Normalization for Image Harmonization [14.77918186672189]
To acquire photo-realistic composite images, one must adjust the appearance and visual style of the foreground to be compatible with the background.
Existing deep learning methods for harmonizing composite images directly learn an image mapping network from the composite to the real one.
We propose a Region-aware Adaptive Instance Normalization (RAIN) module, which explicitly formulates the visual style from the background and adaptively applies them to the foreground.
arXiv Detail & Related papers (2021-06-05T09:57:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.