Highly Efficient Natural Image Matting
- URL: http://arxiv.org/abs/2110.12748v1
- Date: Mon, 25 Oct 2021 09:23:46 GMT
- Title: Highly Efficient Natural Image Matting
- Authors: Yijie Zhong, Bo Li, Lv Tang, Hao Tang, Shouhong Ding
- Abstract summary: We propose a trimap-free natural image matting method with a lightweight model.
We construct an extremely light-weighted model, which achieves comparable performance with 1% (344k) of large models on popular natural image benchmarks.
- Score: 15.977598189574659
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Over the last few years, deep learning based approaches have achieved
outstanding improvements in natural image matting. However, there are still two
drawbacks that impede the widespread application of image matting: the reliance
on user-provided trimaps and the heavy model sizes. In this paper, we propose a
trimap-free natural image matting method with a lightweight model. With a
lightweight basic convolution block, we build a two-stages framework:
Segmentation Network (SN) is designed to capture sufficient semantics and
classify the pixels into unknown, foreground and background regions; Matting
Refine Network (MRN) aims at capturing detailed texture information and
regressing accurate alpha values. With the proposed cross-level fusion Module
(CFM), SN can efficiently utilize multi-scale features with less computational
cost. Efficient non-local attention module (ENA) in MRN can efficiently model
the relevance between different pixels and help regress high-quality alpha
values. Utilizing these techniques, we construct an extremely light-weighted
model, which achieves comparable performance with ~1\% parameters (344k) of
large models on popular natural image matting benchmarks.
Related papers
- Multi-Head Attention Residual Unfolded Network for Model-Based Pansharpening [2.874893537471256]
Unfolding fusion methods integrate the powerful representation capabilities of deep learning with the robustness of model-based approaches.
In this paper, we propose a model-based deep unfolded method for satellite image fusion.
Experimental results on PRISMA, Quickbird, and WorldView2 datasets demonstrate the superior performance of our method.
arXiv Detail & Related papers (2024-09-04T13:05:00Z) - Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - Improving Pixel-based MIM by Reducing Wasted Modeling Capability [77.99468514275185]
We propose a new method that explicitly utilizes low-level features from shallow layers to aid pixel reconstruction.
To the best of our knowledge, we are the first to systematically investigate multi-level feature fusion for isotropic architectures.
Our method yields significant performance gains, such as 1.2% on fine-tuning, 2.8% on linear probing, and 2.6% on semantic segmentation.
arXiv Detail & Related papers (2023-08-01T03:44:56Z) - Efficient Context Integration through Factorized Pyramidal Learning for
Ultra-Lightweight Semantic Segmentation [1.0499611180329804]
We propose a novel Factorized Pyramidal Learning (FPL) module to aggregate rich contextual information in an efficient manner.
We decompose the spatial pyramid into two stages which enables a simple and efficient feature fusion within the module to solve the notorious checkerboard effect.
Based on the FPL module and FIR unit, we propose an ultra-lightweight real-time network, called FPLNet, which achieves state-of-the-art accuracy-efficiency trade-off.
arXiv Detail & Related papers (2023-02-23T05:34:51Z) - DeepDC: Deep Distance Correlation as a Perceptual Image Quality
Evaluator [53.57431705309919]
ImageNet pre-trained deep neural networks (DNNs) show notable transferability for building effective image quality assessment (IQA) models.
We develop a novel full-reference IQA (FR-IQA) model based exclusively on pre-trained DNN features.
We conduct comprehensive experiments to demonstrate the superiority of the proposed quality model on five standard IQA datasets.
arXiv Detail & Related papers (2022-11-09T14:57:27Z) - Spatial-Separated Curve Rendering Network for Efficient and
High-Resolution Image Harmonization [59.19214040221055]
We propose a novel spatial-separated curve rendering network (S$2$CRNet) for efficient and high-resolution image harmonization.
The proposed method reduces more than 90% parameters compared with previous methods.
Our method can work smoothly on higher resolution images in real-time which is more than 10$times$ faster than the existing methods.
arXiv Detail & Related papers (2021-09-13T07:20:16Z) - TextureWGAN: Texture Preserving WGAN with MLE Regularizer for Inverse
Problems [4.112614964808275]
Among all proposed methods, the most popular and effective method is the convolutional neural network (CNN) with mean square error (MSE)
We proposed a new method based on Wasserstein GAN (WGAN) for inverse problems.
We showed that the WGAN-based method was effective to preserve image texture. It also used a maximum likelihood estimation (MLE) regularizer to preserve pixel fidelity.
arXiv Detail & Related papers (2020-08-11T17:06:34Z) - FeatMatch: Feature-Based Augmentation for Semi-Supervised Learning [64.32306537419498]
We propose a novel learned feature-based refinement and augmentation method that produces a varied set of complex transformations.
These transformations also use information from both within-class and across-class representations that we extract through clustering.
We demonstrate that our method is comparable to current state of art for smaller datasets while being able to scale up to larger datasets.
arXiv Detail & Related papers (2020-07-16T17:55:31Z) - Locally Masked Convolution for Autoregressive Models [107.4635841204146]
LMConv is a simple modification to the standard 2D convolution that allows arbitrary masks to be applied to the weights at each location in the image.
We learn an ensemble of distribution estimators that share parameters but differ in generation order, achieving improved performance on whole-image density estimation.
arXiv Detail & Related papers (2020-06-22T17:59:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.