Fast Soft Color Segmentation
- URL: http://arxiv.org/abs/2004.08096v1
- Date: Fri, 17 Apr 2020 07:43:33 GMT
- Title: Fast Soft Color Segmentation
- Authors: Naofumi Akimoto, Huachun Zhu, Yanghua Jin, Yoshimitsu Aoki
- Abstract summary: We address the problem of soft color segmentation, defined as decomposing a given image into several RGBA layers.
We propose a neural network based method for this task that decomposes a given image into multiple layers in a single forward pass.
Our method achieves promising quality without existing issue of inference speed for iterative approaches.
- Score: 10.154836127889487
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We address the problem of soft color segmentation, defined as decomposing a
given image into several RGBA layers, each containing only homogeneous color
regions. The resulting layers from decomposition pave the way for applications
that benefit from layer-based editing, such as recoloring and compositing of
images and videos. The current state-of-the-art approach for this problem is
hindered by slow processing time due to its iterative nature, and consequently
does not scale to certain real-world scenarios. To address this issue, we
propose a neural network based method for this task that decomposes a given
image into multiple layers in a single forward pass. Furthermore, our method
separately decomposes the color layers and the alpha channel layers. By
leveraging a novel training objective, our method achieves proper assignment of
colors amongst layers. As a consequence, our method achieve promising quality
without existing issue of inference speed for iterative approaches. Our
thorough experimental analysis shows that our method produces qualitative and
quantitative results comparable to previous methods while achieving a 300,000x
speed improvement. Finally, we utilize our proposed method on several
applications, and demonstrate its speed advantage, especially in video editing.
Related papers
- Efficient Diffusion Model for Image Restoration by Residual Shifting [63.02725947015132]
This study proposes a novel and efficient diffusion model for image restoration.
Our method avoids the need for post-acceleration during inference, thereby avoiding the associated performance deterioration.
Our method achieves superior or comparable performance to current state-of-the-art methods on three classical IR tasks.
arXiv Detail & Related papers (2024-03-12T05:06:07Z) - Control Color: Multimodal Diffusion-based Interactive Image Colorization [81.68817300796644]
Control Color (Ctrl Color) is a multi-modal colorization method that leverages the pre-trained Stable Diffusion (SD) model.
We present an effective way to encode user strokes to enable precise local color manipulation.
We also introduce a novel module based on self-attention and a content-guided deformable autoencoder to address the long-standing issues of color overflow and inaccurate coloring.
arXiv Detail & Related papers (2024-02-16T17:51:13Z) - Stain-invariant self supervised learning for histopathology image
analysis [74.98663573628743]
We present a self-supervised algorithm for several classification tasks within hematoxylin and eosin stained images of breast cancer.
Our method achieves the state-of-the-art performance on several publicly available breast cancer datasets.
arXiv Detail & Related papers (2022-11-14T18:16:36Z) - ParaColorizer: Realistic Image Colorization using Parallel Generative
Networks [1.7778609937758327]
Grayscale image colorization is a fascinating application of AI for information restoration.
We present a parallel GAN-based colorization framework.
We show the shortcomings of the non-perceptual evaluation metrics commonly used to assess multi-modal problems.
arXiv Detail & Related papers (2022-08-17T13:49:44Z) - Saliency Constrained Arbitrary Image Style Transfer using SIFT and DCNN [22.57205921266602]
When common neural style transfer methods are used, the textures and colors in the style image are usually transferred imperfectly to the content image.
This paper proposes a novel saliency constrained method to reduce or avoid such effects.
The experiments show that the saliency maps of source images can help find the correct matching and avoid artifacts.
arXiv Detail & Related papers (2022-01-14T09:00:55Z) - Temporally Consistent Video Colorization with Deep Feature Propagation
and Self-regularization Learning [90.38674162878496]
We propose a novel temporally consistent video colorization framework (TCVC)
TCVC effectively propagates frame-level deep features in a bidirectional way to enhance the temporal consistency of colorization.
Experiments demonstrate that our method can not only obtain visually pleasing colorized video, but also achieve clearly better temporal consistency than state-of-the-art methods.
arXiv Detail & Related papers (2021-10-09T13:00:14Z) - TUCaN: Progressively Teaching Colourisation to Capsules [13.50327471049997]
We introduce a novel downsampling upsampling architecture named TUCaN (Tiny UCapsNet)
We pose the problem as a per pixel colour classification task that identifies colours as a bin in a quantized space.
To train the network, in contrast with the standard end to end learning method, we propose the progressive learning scheme to extract the context of objects.
arXiv Detail & Related papers (2021-06-29T08:44:15Z) - Fast, Self Supervised, Fully Convolutional Color Normalization of H&E
Stained Images [3.1329883315045106]
Color variation causes problems for the deployment of deep learning-based solutions for automatic diagnosis system in histopathology.
We propose a color normalization technique, which is fast during its self-supervised training as well as inference.
Our method is based on a lightweight fully-convolutional neural network and can be easily attached to a deep learning-based pipeline as a pre-processing block.
arXiv Detail & Related papers (2020-11-30T17:05:58Z) - Deep Image Compositing [93.75358242750752]
We propose a new method which can automatically generate high-quality image composites without any user input.
Inspired by Laplacian pyramid blending, a dense-connected multi-stream fusion network is proposed to effectively fuse the information from the foreground and background images.
Experiments show that the proposed method can automatically generate high-quality composites and outperforms existing methods both qualitatively and quantitatively.
arXiv Detail & Related papers (2020-11-04T06:12:24Z) - Fast Multi-Level Foreground Estimation [0.4588028371034407]
The resulting alpha matte describes pixel-wise to what amount foreground and background colors contribute to the color of the composite image.
While most methods in literature focus on estimating the alpha matte, the process of estimating the foreground colors given the input image and its alpha matte is often neglected.
We propose a novel method for foreground estimation given the alpha matte.
arXiv Detail & Related papers (2020-06-26T13:16:13Z) - Instance-aware Image Colorization [51.12040118366072]
In this paper, we propose a method for achieving instance-aware colorization.
Our network architecture leverages an off-the-shelf object detector to obtain cropped object images.
We use a similar network to extract the full-image features and apply a fusion module to predict the final colors.
arXiv Detail & Related papers (2020-05-21T17:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.