IICNet: A Generic Framework for Reversible Image Conversion
- URL: http://arxiv.org/abs/2109.04242v1
- Date: Thu, 9 Sep 2021 13:06:59 GMT
- Title: IICNet: A Generic Framework for Reversible Image Conversion
- Authors: Ka Leong Cheng and Yueqi Xie and Qifeng Chen
- Abstract summary: Reversible image conversion (RIC) aims to build a reversible transformation between specific visual content (e.g., short videos) and an embedding image.
This work develops Invertible Image Conversion Net (IICNet) as a generic solution to various RIC tasks due to its strong capacity and task-independent design.
- Score: 40.21904131503064
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reversible image conversion (RIC) aims to build a reversible transformation
between specific visual content (e.g., short videos) and an embedding image,
where the original content can be restored from the embedding when necessary.
This work develops Invertible Image Conversion Net (IICNet) as a generic
solution to various RIC tasks due to its strong capacity and task-independent
design. Unlike previous encoder-decoder based methods, IICNet maintains a
highly invertible structure based on invertible neural networks (INNs) to
better preserve the information during conversion. We use a relation module and
a channel squeeze layer to improve the INN nonlinearity to extract cross-image
relations and the network flexibility, respectively. Experimental results
demonstrate that IICNet outperforms the specifically-designed methods on
existing RIC tasks and can generalize well to various newly-explored tasks.
With our generic IICNet, we no longer need to hand-engineer task-specific
embedding networks for rapidly occurring visual content. Our source codes are
available at: https://github.com/felixcheng97/IICNet.
Related papers
- Multiscale Low-Frequency Memory Network for Improved Feature Extraction
in Convolutional Neural Networks [13.815116154370834]
We introduce a novel framework, the Multiscale Low-Frequency Memory (MLFM) Network.
The MLFM efficiently preserves low-frequency information, enhancing performance in targeted computer vision tasks.
Our work builds upon the existing CNN foundations and paves the way for future advancements in computer vision.
arXiv Detail & Related papers (2024-03-13T00:48:41Z) - Adaptive Convolutional Neural Network for Image Super-resolution [43.06377001247278]
We propose a adaptive convolutional neural network for image super-resolution (ADSRNet)
The upper network can enhance relation of context information, salient information relation of a kernel mapping and relations of shallow and deep layers.
The lower network utilizes a symmetric architecture to enhance relations of different layers to mine more structural information.
arXiv Detail & Related papers (2024-02-24T03:44:06Z) - Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - TcGAN: Semantic-Aware and Structure-Preserved GANs with Individual
Vision Transformer for Fast Arbitrary One-Shot Image Generation [11.207512995742999]
One-shot image generation (OSG) with generative adversarial networks that learn from the internal patches of a given image has attracted world wide attention.
We propose a novel structure-preserved method TcGAN with individual vision transformer to overcome the shortcomings of the existing one-shot image generation methods.
arXiv Detail & Related papers (2023-02-16T03:05:59Z) - Cross-receptive Focused Inference Network for Lightweight Image
Super-Resolution [64.25751738088015]
Transformer-based methods have shown impressive performance in single image super-resolution (SISR) tasks.
Transformers that need to incorporate contextual information to extract features dynamically are neglected.
We propose a lightweight Cross-receptive Focused Inference Network (CFIN) that consists of a cascade of CT Blocks mixed with CNN and Transformer.
arXiv Detail & Related papers (2022-07-06T16:32:29Z) - Lightweight Bimodal Network for Single-Image Super-Resolution via
Symmetric CNN and Recursive Transformer [27.51790638626891]
Single-image super-resolution (SISR) has achieved significant breakthroughs with the development of deep learning.
To solve this issue, we propose a Lightweight Bimodal Network (LBNet) for SISR.
Specifically, an effective Symmetric CNN is designed for local feature extraction and coarse image reconstruction.
arXiv Detail & Related papers (2022-04-28T04:43:22Z) - Image Captioning In the Transformer Age [71.06437715212911]
Image Captioning (IC) has achieved astonishing developments by incorporating various techniques into the CNN-RNN encoder-decoder architecture.
This paper analyzes the connections between IC with some popular self-supervised learning paradigms.
arXiv Detail & Related papers (2022-04-15T08:13:39Z) - Learning Deep Interleaved Networks with Asymmetric Co-Attention for
Image Restoration [65.11022516031463]
We present a deep interleaved network (DIN) that learns how information at different states should be combined for high-quality (HQ) images reconstruction.
In this paper, we propose asymmetric co-attention (AsyCA) which is attached at each interleaved node to model the feature dependencies.
Our presented DIN can be trained end-to-end and applied to various image restoration tasks.
arXiv Detail & Related papers (2020-10-29T15:32:00Z) - BiO-Net: Learning Recurrent Bi-directional Connections for
Encoder-Decoder Architecture [82.64881585566825]
We present a novel Bi-directional O-shape network (BiO-Net) that reuses the building blocks in a recurrent manner without introducing any extra parameters.
Our method significantly outperforms the vanilla U-Net as well as other state-of-the-art methods.
arXiv Detail & Related papers (2020-07-01T05:07:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.