WaveMixSR: A Resource-efficient Neural Network for Image
Super-resolution
- URL: http://arxiv.org/abs/2307.00430v1
- Date: Sat, 1 Jul 2023 21:25:03 GMT
- Title: WaveMixSR: A Resource-efficient Neural Network for Image
Super-resolution
- Authors: Pranav Jeevan, Akella Srinidhi, Pasunuri Prathiba, Amit Sethi
- Abstract summary: We propose a new neural network -- WaveMixSR -- for image super-resolution based on WaveMix architecture.
WaveMixSR achieves competitive performance in all datasets and reaches state-of-the-art performance in the BSD100 dataset on multiple super-resolution tasks.
- Score: 2.0477182014909205
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Image super-resolution research recently been dominated by transformer models
which need higher computational resources than CNNs due to the quadratic
complexity of self-attention. We propose a new neural network -- WaveMixSR --
for image super-resolution based on WaveMix architecture which uses a
2D-discrete wavelet transform for spatial token-mixing. Unlike
transformer-based models, WaveMixSR does not unroll the image as a sequence of
pixels/patches. It uses the inductive bias of convolutions along with the
lossless token-mixing property of wavelet transform to achieve higher
performance while requiring fewer resources and training data. We compare the
performance of our network with other state-of-the-art methods for image
super-resolution. Our experiments show that WaveMixSR achieves competitive
performance in all datasets and reaches state-of-the-art performance in the
BSD100 dataset on multiple super-resolution tasks. Our model is able to achieve
this performance using less training data and computational resources while
maintaining high parameter efficiency compared to current state-of-the-art
models.
Related papers
- WaveMixSR-V2: Enhancing Super-resolution with Higher Efficiency [4.093503153499691]
We present an enhanced version of the WaveMixSR architecture by replacing the traditional convolution layer with a pixel shuffle operation.
Our experiments demonstrate that our enhanced model -- WaveMixSR-V2 -- outperforms other architectures in multiple super-resolution tasks.
arXiv Detail & Related papers (2024-09-16T04:16:52Z) - DGNet: Dynamic Gradient-Guided Network for Water-Related Optics Image
Enhancement [77.0360085530701]
Underwater image enhancement (UIE) is a challenging task due to the complex degradation caused by underwater environments.
Previous methods often idealize the degradation process, and neglect the impact of medium noise and object motion on the distribution of image features.
Our approach utilizes predicted images to dynamically update pseudo-labels, adding a dynamic gradient to optimize the network's gradient space.
arXiv Detail & Related papers (2023-12-12T06:07:21Z) - WavePaint: Resource-efficient Token-mixer for Self-supervised Inpainting [2.3014300466616078]
This paper diverges from vision transformers by using a computationally-efficient WaveMix-based fully convolutional architecture -- WavePaint.
It uses a 2D-discrete wavelet transform (DWT) for spatial and multi-resolution token-mixing along with convolutional layers.
Our model even outperforms current GAN-based architectures in CelebA-HQ dataset without using an adversarially trainable discriminator.
arXiv Detail & Related papers (2023-07-01T18:41:34Z) - ShuffleMixer: An Efficient ConvNet for Image Super-Resolution [88.86376017828773]
We propose ShuffleMixer, for lightweight image super-resolution that explores large convolution and channel split-shuffle operation.
Specifically, we develop a large depth-wise convolution and two projection layers based on channel splitting and shuffling as the basic component to mix features efficiently.
Experimental results demonstrate that the proposed ShuffleMixer is about 6x smaller than the state-of-the-art methods in terms of model parameters and FLOPs.
arXiv Detail & Related papers (2022-05-30T15:26:52Z) - WaveMix: A Resource-efficient Neural Network for Image Analysis [3.4927288761640565]
WaveMix is resource-efficient and yet generalizable and scalable.
Networks achieve comparable or better accuracy than the state-of-the-art convolutional neural networks.
WaveMix establishes new benchmarks for segmentation on Cityscapes.
arXiv Detail & Related papers (2022-05-28T09:08:50Z) - Hybrid Pixel-Unshuffled Network for Lightweight Image Super-Resolution [64.54162195322246]
Convolutional neural network (CNN) has achieved great success on image super-resolution (SR)
Most deep CNN-based SR models take massive computations to obtain high performance.
We propose a novel Hybrid Pixel-Unshuffled Network (HPUN) by introducing an efficient and effective downsampling module into the SR task.
arXiv Detail & Related papers (2022-03-16T20:10:41Z) - WaveMix: Resource-efficient Token Mixing for Images [2.7188347260210466]
We present WaveMix as an alternative neural architecture that uses a multi-scale 2D discrete wavelet transform (DWT) for spatial token mixing.
WaveMix has achieved State-of-the-art (SOTA) results in EMNIST Byclass and EMNIST Balanced datasets.
arXiv Detail & Related papers (2022-03-07T20:15:17Z) - Restormer: Efficient Transformer for High-Resolution Image Restoration [118.9617735769827]
convolutional neural networks (CNNs) perform well at learning generalizable image priors from large-scale data.
Transformers have shown significant performance gains on natural language and high-level vision tasks.
Our model, named Restoration Transformer (Restormer), achieves state-of-the-art results on several image restoration tasks.
arXiv Detail & Related papers (2021-11-18T18:59:10Z) - Asymmetric CNN for image super-resolution [102.96131810686231]
Deep convolutional neural networks (CNNs) have been widely applied for low-level vision over the past five years.
We propose an asymmetric CNN (ACNet) comprising an asymmetric block (AB), a mem?ory enhancement block (MEB) and a high-frequency feature enhancement block (HFFEB) for image super-resolution.
Our ACNet can effectively address single image super-resolution (SISR), blind SISR and blind SISR of blind noise problems.
arXiv Detail & Related papers (2021-03-25T07:10:46Z) - Learning Frequency-aware Dynamic Network for Efficient Super-Resolution [56.98668484450857]
This paper explores a novel frequency-aware dynamic network for dividing the input into multiple parts according to its coefficients in the discrete cosine transform (DCT) domain.
In practice, the high-frequency part will be processed using expensive operations and the lower-frequency part is assigned with cheap operations to relieve the computation burden.
Experiments conducted on benchmark SISR models and datasets show that the frequency-aware dynamic network can be employed for various SISR neural architectures.
arXiv Detail & Related papers (2021-03-15T12:54:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.