Color2Style: Real-Time Exemplar-Based Image Colorization with
Self-Reference Learning and Deep Feature Modulation
- URL: http://arxiv.org/abs/2106.08017v2
- Date: Wed, 16 Jun 2021 05:46:23 GMT
- Title: Color2Style: Real-Time Exemplar-Based Image Colorization with
Self-Reference Learning and Deep Feature Modulation
- Authors: Hengyuan Zhao, Wenhao Wu, Yihao Liu, Dongliang He
- Abstract summary: We present a deep exemplar-based image colorization approach named Color2Style to resurrect grayscale image media by filling them with vibrant colors.
Our method exploits a simple yet effective deep feature modulation (DFM) module, which injects the color embeddings extracted from the reference image into the deep representations of the input grayscale image.
- Score: 29.270149925368674
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Legacy black-and-white photos are riddled with people's nostalgia and
glorious memories of the past. To better relive the elapsed frozen moments, in
this paper, we present a deep exemplar-based image colorization approach named
Color2Style to resurrect these grayscale image media by filling them with
vibrant colors. Generally, for exemplar-based colorization, unsupervised and
unpaired training are usually adopted, due to the difficulty of obtaining input
and ground truth image pairs. To train an exemplar-based colorization model,
current algorithms usually strive to achieve two procedures: i) retrieving a
large number of reference images with high similarity in advance, which is
inevitably time-consuming and tedious; ii) designing complicated modules to
transfer the colors of the reference image to the grayscale image, by
calculating and leveraging the deep semantic correspondence between them (e.g.,
non-local operation). Contrary to the previous methods, we solve and simplify
the above two steps in one end-to-end learning procedure. First, we adopt a
self-augmented self-reference training scheme, where the reference image is
generated by graphical transformations from the original colorful one whereby
the training can be formulated in a paired manner. Second, instead of computing
complex and inexplicable correspondence maps, our method exploits a simple yet
effective deep feature modulation (DFM) module, which injects the color
embeddings extracted from the reference image into the deep representations of
the input grayscale image. Such design is much more lightweight and
intelligible, achieving appealing performance with real-time processing speed.
Moreover, our model does not require multifarious loss functions and
regularization terms like existing methods, but only two widely used loss
functions. Codes and models will be available at
https://github.com/zhaohengyuan1/Color2Style.
Related papers
- Training-free Color-Style Disentanglement for Constrained Text-to-Image Synthesis [16.634138745034733]
We present the first training-free, test-time-only method to disentangle and condition text-to-image models on color and style attributes from reference image.
arXiv Detail & Related papers (2024-09-04T04:16:58Z) - Automatic Controllable Colorization via Imagination [55.489416987587305]
We propose a framework for automatic colorization that allows for iterative editing and modifications.
By understanding the content within a grayscale image, we utilize a pre-trained image generation model to generate multiple images that contain the same content.
These images serve as references for coloring, mimicking the process of human experts.
arXiv Detail & Related papers (2024-04-08T16:46:07Z) - Control Color: Multimodal Diffusion-based Interactive Image Colorization [81.68817300796644]
Control Color (Ctrl Color) is a multi-modal colorization method that leverages the pre-trained Stable Diffusion (SD) model.
We present an effective way to encode user strokes to enable precise local color manipulation.
We also introduce a novel module based on self-attention and a content-guided deformable autoencoder to address the long-standing issues of color overflow and inaccurate coloring.
arXiv Detail & Related papers (2024-02-16T17:51:13Z) - UniGS: Unified Representation for Image Generation and Segmentation [105.08152635402858]
We use a colormap to represent entity-level masks, addressing the challenge of varying entity numbers.
Two novel modules, including the location-aware color palette and progressive dichotomy module, are proposed to support our mask representation.
arXiv Detail & Related papers (2023-12-04T15:59:27Z) - Incorporating Ensemble and Transfer Learning For An End-To-End
Auto-Colorized Image Detection Model [0.0]
This paper presents a novel approach that combines the advantages of transfer and ensemble learning approaches to help reduce training time and resource requirements.
The proposed model shows promising results, with accuracy ranging from 94.55% to 99.13%.
arXiv Detail & Related papers (2023-09-25T19:22:57Z) - Improved Diffusion-based Image Colorization via Piggybacked Models [19.807766482434563]
We introduce a colorization model piggybacking on the existing powerful T2I diffusion model.
A diffusion guider is designed to incorporate the pre-trained weights of the latent diffusion model.
A lightness-aware VQVAE will then generate the colorized result with pixel-perfect alignment to the given grayscale image.
arXiv Detail & Related papers (2023-04-21T16:23:24Z) - BiSTNet: Semantic Image Prior Guided Bidirectional Temporal Feature
Fusion for Deep Exemplar-based Video Colorization [70.14893481468525]
We present an effective BiSTNet to explore colors of reference exemplars and utilize them to help video colorization.
We first establish the semantic correspondence between each frame and the reference exemplars in deep feature space to explore color information from reference exemplars.
We develop a mixed expert block to extract semantic information for modeling the object boundaries of frames so that the semantic image prior can better guide the colorization process.
arXiv Detail & Related papers (2022-12-05T13:47:15Z) - ParaColorizer: Realistic Image Colorization using Parallel Generative
Networks [1.7778609937758327]
Grayscale image colorization is a fascinating application of AI for information restoration.
We present a parallel GAN-based colorization framework.
We show the shortcomings of the non-perceptual evaluation metrics commonly used to assess multi-modal problems.
arXiv Detail & Related papers (2022-08-17T13:49:44Z) - Detecting Recolored Image by Spatial Correlation [60.08643417333974]
Image recoloring is an emerging editing technique that can manipulate the color values of an image to give it a new style.
In this paper, we explore a solution from the perspective of the spatial correlation, which exhibits the generic detection capability for both conventional and deep learning-based recoloring.
Our method achieves the state-of-the-art detection accuracy on multiple benchmark datasets and exhibits well generalization for unknown types of recoloring methods.
arXiv Detail & Related papers (2022-04-23T01:54:06Z) - Semantic-Sparse Colorization Network for Deep Exemplar-based
Colorization [23.301799487207035]
Exemplar-based colorization approaches rely on reference image to provide plausible colors for target gray-scale image.
We propose Semantic-Sparse Colorization Network (SSCN) to transfer both the global image style and semantic-related colors to the gray-scale image.
Our network can perfectly balance the global and local colors while alleviating the ambiguous matching problem.
arXiv Detail & Related papers (2021-12-02T15:35:10Z) - Instance-aware Image Colorization [51.12040118366072]
In this paper, we propose a method for achieving instance-aware colorization.
Our network architecture leverages an off-the-shelf object detector to obtain cropped object images.
We use a similar network to extract the full-image features and apply a fusion module to predict the final colors.
arXiv Detail & Related papers (2020-05-21T17:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.