Hierarchical Image Peeling: A Flexible Scale-space Filtering Framework
- URL: http://arxiv.org/abs/2104.01534v1
- Date: Sun, 4 Apr 2021 04:08:14 GMT
- Title: Hierarchical Image Peeling: A Flexible Scale-space Filtering Framework
- Authors: Fu Yuanbin and Guoxiaojie and Hu Qiming and Lin Di and Ma Jiayi and
Ling Haibin
- Abstract summary: hierarchical image organization has been witnessed by a wide spectrum of applications in computer vision and graphics.
This work designs a modern framework for disassembling an image into a family of derived signals from a scale-space perspective.
A compact recurrent network, namely hierarchical image peeling net, is customized to efficiently and effectively fulfill the task.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The importance of hierarchical image organization has been witnessed by a
wide spectrum of applications in computer vision and graphics. Different from
image segmentation with the spatial whole-part consideration, this work designs
a modern framework for disassembling an image into a family of derived signals
from a scale-space perspective. Specifically, we first offer a formal
definition of image disassembly. Then, by concerning desired properties, such
as peeling hierarchy and structure preservation, we convert the original
complex problem into a series of two-component separation sub-problems,
significantly reducing the complexity. The proposed framework is flexible to
both supervised and unsupervised settings. A compact recurrent network, namely
hierarchical image peeling net, is customized to efficiently and effectively
fulfill the task, which is about 3.5Mb in size, and can handle 1080p images in
more than 60 fps per recurrence on a GTX 2080Ti GPU, making it attractive for
practical use. Both theoretical findings and experimental results are provided
to demonstrate the efficacy of the proposed framework, reveal its superiority
over other state-of-the-art alternatives, and show its potential to various
applicable scenarios. Our code is available at
\url{https://github.com/ForawardStar/HIPe}.
Related papers
- Mixed Hierarchy Network for Image Restoration [0.0]
We present a mixed hierarchy network that can balance quality and system complexity in image restoration.
Our model first learns the contextual information using encoder-decoder architectures, and then combines them with high-resolution branches that preserve spatial detail.
The resulting tightly interlinked hierarchy architecture, named as MHNet, delivers strong performance gains on several image restoration tasks.
arXiv Detail & Related papers (2023-02-19T12:18:45Z) - High-Quality Pluralistic Image Completion via Code Shared VQGAN [51.7805154545948]
We present a novel framework for pluralistic image completion that can achieve both high quality and diversity at much faster inference speed.
Our framework is able to learn semantically-rich discrete codes efficiently and robustly, resulting in much better image reconstruction quality.
arXiv Detail & Related papers (2022-04-05T01:47:35Z) - HIPA: Hierarchical Patch Transformer for Single Image Super Resolution [62.7081074931892]
This paper presents HIPA, a novel Transformer architecture that progressively recovers the high resolution image using a hierarchical patch partition.
We build a cascaded model that processes an input image in multiple stages, where we start with tokens with small patch sizes and gradually merge to the full resolution.
Such a hierarchical patch mechanism not only explicitly enables feature aggregation at multiple resolutions but also adaptively learns patch-aware features for different image regions.
arXiv Detail & Related papers (2022-03-19T05:09:34Z) - Leveraging Image Complexity in Macro-Level Neural Network Design for
Medical Image Segmentation [3.974175960216864]
We show that image complexity can be used as a guideline in choosing what is best for a given dataset.
For high-complexity datasets, a shallow network running on the original images may yield better segmentation results than a deep network running on downsampled images.
arXiv Detail & Related papers (2021-12-21T09:49:47Z) - Spatially-Adaptive Image Restoration using Distortion-Guided Networks [51.89245800461537]
We present a learning-based solution for restoring images suffering from spatially-varying degradations.
We propose SPAIR, a network design that harnesses distortion-localization information and dynamically adjusts to difficult regions in the image.
arXiv Detail & Related papers (2021-08-19T11:02:25Z) - Generating Diverse Structure for Image Inpainting With Hierarchical
VQ-VAE [74.29384873537587]
We propose a two-stage model for diverse inpainting, where the first stage generates multiple coarse results each of which has a different structure, and the second stage refines each coarse result separately by augmenting texture.
Experimental results on CelebA-HQ, Places2, and ImageNet datasets show that our method not only enhances the diversity of the inpainting solutions but also improves the visual quality of the generated multiple images.
arXiv Detail & Related papers (2021-03-18T05:10:49Z) - Searching for Controllable Image Restoration Networks [57.23583915884236]
Existing methods require separate inference through the entire network per each output.
We propose a novel framework based on a neural architecture search technique that enables efficient generation of multiple imagery effects.
arXiv Detail & Related papers (2020-12-21T10:08:18Z) - Conceptual Compression via Deep Structure and Texture Synthesis [42.68994438290913]
We propose a novel conceptual compression framework that encodes visual data into compact structure and texture representations, then decodes in a deep synthesis fashion.
In particular, we propose to compress images by a dual-layered model consisting of two complementary visual features.
At the encoder side, the structural maps and texture representations are individually extracted and compressed, generating the compact, interpretable, inter-operable bitstreams.
During the decoding stage, a hierarchical fusion GAN (HF-GAN) is proposed to learn the synthesis paradigm where the textures are rendered into the decoded structural maps, leading to high-quality reconstruction
arXiv Detail & Related papers (2020-11-10T08:48:32Z) - TSIT: A Simple and Versatile Framework for Image-to-Image Translation [103.92203013154403]
We introduce a simple and versatile framework for image-to-image translation.
We provide a carefully designed two-stream generative model with newly proposed feature transformations.
This allows multi-scale semantic structure information and style representation to be effectively captured and fused by the network.
A systematic study compares the proposed method with several state-of-the-art task-specific baselines, verifying its effectiveness in both perceptual quality and quantitative evaluations.
arXiv Detail & Related papers (2020-07-23T15:34:06Z) - Contextual Residual Aggregation for Ultra High-Resolution Image
Inpainting [12.839962012888199]
We propose a Contextual Residual Aggregation (CRA) mechanism that can produce high-frequency residuals for missing contents.
CRA mechanism produces high-frequency residuals for missing contents by weighted aggregating residuals from contextual patches.
We train the proposed model on small images with resolutions 512x512 and perform inference on high-resolution images, achieving compelling inpainting quality.
arXiv Detail & Related papers (2020-05-19T18:55:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.