AlphaNet: An Attention Guided Deep Network for Automatic Image Matting
- URL: http://arxiv.org/abs/2003.03613v1
- Date: Sat, 7 Mar 2020 17:25:21 GMT
- Title: AlphaNet: An Attention Guided Deep Network for Automatic Image Matting
- Authors: Rishab Sharma, Rahul Deora and Anirudha Vishvakarma
- Abstract summary: We propose an end to end solution for image matting i.e. high-precision extraction of foreground objects from natural images.
We propose a method that assimilates semantic segmentation and deep image matting processes into a single network to generate semantic mattes.
We also construct a fashion e-commerce focused dataset with high-quality alpha mattes to facilitate the training and evaluation for image matting.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose an end to end solution for image matting i.e
high-precision extraction of foreground objects from natural images. Image
matting and background detection can be achieved easily through chroma keying
in a studio setting when the background is either pure green or blue.
Nonetheless, image matting in natural scenes with complex and uneven depth
backgrounds remains a tedious task that requires human intervention. To achieve
complete automatic foreground extraction in natural scenes, we propose a method
that assimilates semantic segmentation and deep image matting processes into a
single network to generate detailed semantic mattes for image composition task.
The contribution of our proposed method is two-fold, firstly it can be
interpreted as a fully automated semantic image matting method and secondly as
a refinement of existing semantic segmentation models. We propose a novel model
architecture as a combination of segmentation and matting that unifies the
function of upsampling and downsampling operators with the notion of attention.
As shown in our work, attention guided downsampling and upsampling can extract
high-quality boundary details, unlike other normal downsampling and upsampling
techniques. For achieving the same, we utilized an attention guided
encoder-decoder framework which does unsupervised learning for generating an
attention map adaptively from the data to serve and direct the upsampling and
downsampling operators. We also construct a fashion e-commerce focused dataset
with high-quality alpha mattes to facilitate the training and evaluation for
image matting.
Related papers
- Towards Natural Image Matting in the Wild via Real-Scenario Prior [69.96414467916863]
We propose a new matting dataset based on the COCO dataset, namely COCO-Matting.
The built COCO-Matting comprises an extensive collection of 38,251 human instance-level alpha mattes in complex natural scenarios.
For network architecture, the proposed feature-aligned transformer learns to extract fine-grained edge and transparency features.
The proposed matte-aligned decoder aims to segment matting-specific objects and convert coarse masks into high-precision mattes.
arXiv Detail & Related papers (2024-10-09T06:43:19Z) - DiffUHaul: A Training-Free Method for Object Dragging in Images [78.93531472479202]
We propose a training-free method, dubbed DiffUHaul, for the object dragging task.
We first apply attention masking in each denoising step to make the generation more disentangled across different objects.
In the early denoising steps, we interpolate the attention features between source and target images to smoothly fuse new layouts with the original appearance.
arXiv Detail & Related papers (2024-06-03T17:59:53Z) - EFormer: Enhanced Transformer towards Semantic-Contour Features of
Foreground for Portraits Matting [6.468859319728341]
We propose EFormer to enhance the model's attention towards both of the low-frequency semantic and high-frequency contour features.
We build a semantic and contour detector (SCD) to accurately capture both of the low-frequency semantic and high-frequency contour features.
And we design contour-edge extraction branch and semantic extraction branch to extract refined high-frequency contour features and complete low-frequency semantic information.
arXiv Detail & Related papers (2023-08-24T14:45:03Z) - Deep Image Matting: A Comprehensive Survey [85.77905619102802]
This paper presents a review of recent advancements in image matting in the era of deep learning.
We focus on two fundamental sub-tasks: auxiliary input-based image matting and automatic image matting.
We discuss relevant applications of image matting and highlight existing challenges and potential opportunities for future research.
arXiv Detail & Related papers (2023-04-10T15:48:55Z) - Disentangled Pre-training for Image Matting [74.10407744483526]
Image matting requires high-quality pixel-level human annotations to support the training of a deep model.
We propose a self-supervised pre-training approach that can leverage infinite numbers of data to boost the matting performance.
arXiv Detail & Related papers (2023-04-03T08:16:02Z) - Human Perception Modeling for Automatic Natural Image Matting [2.179313476241343]
Natural image matting aims to precisely separate foreground objects from background using alpha matte.
We propose an intuitively-designed trimap-free two-stage matting approach without additional annotations.
Our matting algorithm has competitive performance with current state-of-the-art methods in both trimap-free and trimap-needed aspects.
arXiv Detail & Related papers (2021-03-31T12:08:28Z) - Deep Image Compositing [93.75358242750752]
We propose a new method which can automatically generate high-quality image composites without any user input.
Inspired by Laplacian pyramid blending, a dense-connected multi-stream fusion network is proposed to effectively fuse the information from the foreground and background images.
Experiments show that the proposed method can automatically generate high-quality composites and outperforms existing methods both qualitatively and quantitatively.
arXiv Detail & Related papers (2020-11-04T06:12:24Z) - Bridging Composite and Real: Towards End-to-end Deep Image Matting [88.79857806542006]
We study the roles of semantics and details for image matting.
We propose a novel Glance and Focus Matting network (GFM), which employs a shared encoder and two separate decoders.
Comprehensive empirical studies have demonstrated that GFM outperforms state-of-the-art methods.
arXiv Detail & Related papers (2020-10-30T10:57:13Z) - Natural Image Matting via Guided Contextual Attention [18.034160025888056]
We develop a novel end-to-end approach for natural image matting with a guided contextual attention module.
The proposed method can mimic information flow of affinity-based methods and utilize rich features learned by deep neural networks simultaneously.
Experiment results on Composition-1k testing set and alphamatting.com benchmark dataset demonstrate that our method outperforms state-of-the-art approaches in natural image matting.
arXiv Detail & Related papers (2020-01-13T05:59:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.