Hierarchical and Progressive Image Matting
- URL: http://arxiv.org/abs/2210.06906v1
- Date: Thu, 13 Oct 2022 11:16:49 GMT
- Title: Hierarchical and Progressive Image Matting
- Authors: Yu Qiao, Yuhao Liu, Ziqi Wei, Yuxin Wang, Qiang Cai, Guofeng Zhang,
Xin Yang
- Abstract summary: We propose an end-to-end Hierarchical and Progressive Attention Matting Network (HAttMatting++)
It can better predict the opacity of the foreground from single RGB images without additional input.
We construct a large-scale and challenging image matting dataset comprised of 59, 600 training images and 1000 test images.
- Score: 40.291998690687514
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most matting researches resort to advanced semantics to achieve high-quality
alpha mattes, and direct low-level features combination is usually explored to
complement alpha details. However, we argue that appearance-agnostic
integration can only provide biased foreground details and alpha mattes require
different-level feature aggregation for better pixel-wise opacity perception.
In this paper, we propose an end-to-end Hierarchical and Progressive Attention
Matting Network (HAttMatting++), which can better predict the opacity of the
foreground from single RGB images without additional input. Specifically, we
utilize channel-wise attention to distill pyramidal features and employ spatial
attention at different levels to filter appearance cues. This progressive
attention mechanism can estimate alpha mattes from adaptive semantics and
semantics-indicated boundaries. We also introduce a hybrid loss function fusing
Structural SIMilarity (SSIM), Mean Square Error (MSE), Adversarial loss, and
sentry supervision to guide the network to further improve the overall
foreground structure. Besides, we construct a large-scale and challenging image
matting dataset comprised of 59, 600 training images and 1000 test images (a
total of 646 distinct foreground alpha mattes), which can further improve the
robustness of our hierarchical and progressive aggregation model. Extensive
experiments demonstrate that the proposed HAttMatting++ can capture
sophisticated foreground structures and achieve state-of-the-art performance
with single RGB images as input.
Related papers
- Multi-Head Attention Residual Unfolded Network for Model-Based Pansharpening [2.874893537471256]
Unfolding fusion methods integrate the powerful representation capabilities of deep learning with the robustness of model-based approaches.
In this paper, we propose a model-based deep unfolded method for satellite image fusion.
Experimental results on PRISMA, Quickbird, and WorldView2 datasets demonstrate the superior performance of our method.
arXiv Detail & Related papers (2024-09-04T13:05:00Z) - Rank-Enhanced Low-Dimensional Convolution Set for Hyperspectral Image
Denoising [50.039949798156826]
This paper tackles the challenging problem of hyperspectral (HS) image denoising.
We propose rank-enhanced low-dimensional convolution set (Re-ConvSet)
We then incorporate Re-ConvSet into the widely-used U-Net architecture to construct an HS image denoising method.
arXiv Detail & Related papers (2022-07-09T13:35:12Z) - Prior-Induced Information Alignment for Image Matting [28.90998570043986]
We propose a novel network named Prior-Induced Information Alignment Matting Network (PIIAMatting)
It can efficiently model the distinction of pixel-wise response maps and the correlation of layer-wise feature maps.
PIIAMatting performs favourably against state-of-the-art image matting methods on the Alphamatting.com, Composition-1K and Distinctions-646 dataset.
arXiv Detail & Related papers (2021-06-28T07:46:59Z) - High-resolution Depth Maps Imaging via Attention-based Hierarchical
Multi-modal Fusion [84.24973877109181]
We propose a novel attention-based hierarchical multi-modal fusion network for guided DSR.
We show that our approach outperforms state-of-the-art methods in terms of reconstruction accuracy, running speed and memory efficiency.
arXiv Detail & Related papers (2021-04-04T03:28:33Z) - Smart Scribbles for Image Mating [90.18035889903909]
We propose an interactive framework, referred to as smart scribbles, to guide users to draw few scribbles on the input images.
It infers the most informative regions of an image for drawing scribbles to indicate different categories.
It then spreads these scribbles to the rest of the image via our well-designed two-phase propagation.
arXiv Detail & Related papers (2021-03-31T13:30:49Z) - Deep Burst Super-Resolution [165.90445859851448]
We propose a novel architecture for the burst super-resolution task.
Our network takes multiple noisy RAW images as input, and generates a denoised, super-resolved RGB image as output.
In order to enable training and evaluation on real-world data, we additionally introduce the BurstSR dataset.
arXiv Detail & Related papers (2021-01-26T18:57:21Z) - Multi-scale Information Assembly for Image Matting [35.43994064645042]
We propose a multi-scale information assembly framework (MSIA-matte) to pull out high-quality alpha mattes from single RGB images.
We can achieve state-of-the-art performance compared to most existing matting networks.
arXiv Detail & Related papers (2021-01-07T06:15:48Z) - Bridging Composite and Real: Towards End-to-end Deep Image Matting [88.79857806542006]
We study the roles of semantics and details for image matting.
We propose a novel Glance and Focus Matting network (GFM), which employs a shared encoder and two separate decoders.
Comprehensive empirical studies have demonstrated that GFM outperforms state-of-the-art methods.
arXiv Detail & Related papers (2020-10-30T10:57:13Z) - Hierarchical Opacity Propagation for Image Matting [15.265494938937737]
A novel structure for more direct alpha matte propagation between pixels is in demand.
HOP matting is capable of outperforming state-of-the-art matting methods.
arXiv Detail & Related papers (2020-04-07T10:39:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.