Reconstructed Convolution Module Based Look-Up Tables for Efficient
Image Super-Resolution
- URL: http://arxiv.org/abs/2307.08544v1
- Date: Mon, 17 Jul 2023 15:04:00 GMT
- Title: Reconstructed Convolution Module Based Look-Up Tables for Efficient
Image Super-Resolution
- Authors: Guandu Liu, Yukang Ding, Mading Li, Ming Sun, Xing Wen and Bin Wang
- Abstract summary: Look-up table(LUT)-based methods have shown the great efficacy in single image super-resolution (SR) task.
Previous methods ignore the essential reason of restricted receptive field (RF) size in LUT.
We propose a novel Reconstructed Convolution(RC) module, which decouples channel-wise and spatial calculation.
- Score: 9.715421499605934
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Look-up table(LUT)-based methods have shown the great efficacy in single
image super-resolution (SR) task. However, previous methods ignore the
essential reason of restricted receptive field (RF) size in LUT, which is
caused by the interaction of space and channel features in vanilla convolution.
They can only increase the RF at the cost of linearly increasing LUT size. To
enlarge RF with contained LUT sizes, we propose a novel Reconstructed
Convolution(RC) module, which decouples channel-wise and spatial calculation.
It can be formulated as $n^2$ 1D LUTs to maintain $n\times n$ receptive field,
which is obviously smaller than $n\times n$D LUT formulated before. The LUT
generated by our RC module reaches less than 1/10000 storage compared with
SR-LUT baseline. The proposed Reconstructed Convolution module based LUT
method, termed as RCLUT, can enlarge the RF size by 9 times than the
state-of-the-art LUT-based SR method and achieve superior performance on five
popular benchmark dataset. Moreover, the efficient and robust RC module can be
used as a plugin to improve other LUT-based SR methods. The code is available
at https://github.com/liuguandu/RC-LUT.
Related papers
- ShiftLUT: Spatial Shift Enhanced Look-Up Tables for Efficient Image Restoration [8.845117852325997]
ShiftLUT is a novel framework that attains the largest receptive field among all LUT-based methods while maintaining high efficiency.<n>Compared to the previous state-of-the-art method TinyLUT, ShiftLUT achieves a 3.8$times$ larger receptive field and improves an average PSNR by over 0.21 dB.
arXiv Detail & Related papers (2026-03-01T04:00:23Z) - ALLMod: Exploring $\underline{\mathbf{A}}$rea-Efficiency of $\underline{\mathbf{L}}$UT-based $\underline{\mathbf{L}}$arge Number $\underline{\mathbf{Mod}}$ular Reduction via Hybrid Workloads [18.634794494170617]
High-bit-width operations are crucial for enhancing security.
They are computationally intensive due to the large number of modular operations required.
AllMod is a novel approach that improves the area efficiency of LUT-based large-number modular reduction.
arXiv Detail & Related papers (2025-03-20T07:47:34Z) - ReducedLUT: Table Decomposition with "Don't Care" Conditions [2.2061683015812026]
This paper introduces ReducedLUT, a novel method to reduce the footprint of the LUTs by injecting don't cares into the compression process.
We demonstrate a particular application to machine learning; by replacing unobserved patterns within the training data of neural network models with don't cares, we enable greater compression with minimal model accuracy degradation.
arXiv Detail & Related papers (2024-12-24T18:11:01Z) - LUT Tensor Core: A Software-Hardware Co-Design for LUT-Based Low-Bit LLM Inference [10.608817382813786]
Mixed-precision matrix (mpGEMM) is an important yet underexplored operation involving the multiplication of lower-precision weights with higher-precision activations.<n>Off-the-shelf hardware does not support this operation, leading to indirect, thus inefficient, dequantization-based implementations.<n>We study the lookup table (LUT)-based approach for mpGEMM and find that a conventional LUT implementation fails to achieve the promised gains.
arXiv Detail & Related papers (2024-08-12T08:52:14Z) - PolyLUT-Add: FPGA-based LUT Inference with Wide Inputs [1.730979251211628]
This work introduces PolyLUT-Add, a technique that enhances neuron connectivity by combining $A$ PolyLUT sub-neurons via addition to improve accuracy.
We evaluate our implementation over the MNIST, Jet Substructure classification, and Network Intrusion Detection benchmark and found that for similar accuracy, PolyLUT-Add achieves a LUT reduction of $2.0-13.9times$ with a $1.2-1.6times$ decrease in latency.
arXiv Detail & Related papers (2024-06-07T13:00:57Z) - NAFRSSR: a Lightweight Recursive Network for Efficient Stereo Image Super-Resolution [17.025789865392444]
Stereo image super-resolution (SR) refers to the reconstruction of a high-resolution (HR) image from a pair of low-resolution (LR) images as typically captured by a dual-camera device.
Here, we propose a simple yet efficient stereo image SR model called NAFRSSR.
Our model is composed of nonlinear activation free and group convolution-based blocks (NAFGCBlocks) and depth-separated stereo cross attention modules (DSSCAMs)
arXiv Detail & Related papers (2024-05-14T08:33:40Z) - Frequency-Assisted Mamba for Remote Sensing Image Super-Resolution [49.902047563260496]
We develop the first attempt to integrate the Vision State Space Model (Mamba) for remote sensing image (RSI) super-resolution.
To achieve better SR reconstruction, building upon Mamba, we devise a Frequency-assisted Mamba framework, dubbed FMSR.
Our FMSR features a multi-level fusion architecture equipped with the Frequency Selection Module (FSM), Vision State Space Module (VSSM), and Hybrid Gate Module (HGM)
arXiv Detail & Related papers (2024-05-08T11:09:24Z) - HiRE: High Recall Approximate Top-$k$ Estimation for Efficient LLM
Inference [68.59839755875252]
HiRE comprises of two novel components: (i) a compression scheme to cheaply predict top-$k$ rows/columns with high recall, followed by full computation restricted to the predicted subset, and (ii) DA-TOP-$k$: an efficient multi-device approximate top-$k$ operator.
We demonstrate that on a one billion parameter model, HiRE applied to both the softmax as well as feedforward layers, achieves almost matching pretraining and downstream accuracy, and speeds up inference latency by $1.47times$ on a single TPUv5e device.
arXiv Detail & Related papers (2024-02-14T18:04:36Z) - Transforming Image Super-Resolution: A ConvFormer-based Efficient Approach [58.57026686186709]
We introduce the Convolutional Transformer layer (ConvFormer) and propose a ConvFormer-based Super-Resolution network (CFSR)
CFSR inherits the advantages of both convolution-based and transformer-based approaches.
Experiments demonstrate that CFSR strikes an optimal balance between computational cost and performance.
arXiv Detail & Related papers (2024-01-11T03:08:00Z) - Hundred-Kilobyte Lookup Tables for Efficient Single-Image Super-Resolution [7.403264755337134]
Super-resolution (SR) schemes make heavy use of convolutional neural networks (CNNs), which involve intensive multiply-accumulate (MAC) operations.
This contradicts the regime of edge AI that often runs on devices strained by power, computing, and storage resources.
This work tackles this storage hurdle and innovates hundred-kilobyte LUT (HKLUT) models to amenable to on-chip cache.
arXiv Detail & Related papers (2023-12-11T04:07:34Z) - Instant Complexity Reduction in CNNs using Locality-Sensitive Hashing [50.79602839359522]
We propose HASTE (Hashing for Tractable Efficiency), a parameter-free and data-free module that acts as a plug-and-play replacement for any regular convolution module.
We are able to drastically compress latent feature maps without sacrificing much accuracy by using locality-sensitive hashing (LSH)
In particular, we are able to instantly drop 46.72% of FLOPs while only losing 1.25% accuracy by just swapping the convolution modules in a ResNet34 on CIFAR-10 for our HASTE module.
arXiv Detail & Related papers (2023-09-29T13:09:40Z) - Toward DNN of LUTs: Learning Efficient Image Restoration with Multiple
Look-Up Tables [47.15181829317732]
High-definition screens on edge devices stimulate a strong demand for efficient image restoration algorithms.
The size of a single look-up table grows exponentially with the increase of its indexing capacity.
We propose a universal method to construct multiple LUTs like a neural network, termed MuLUT.
arXiv Detail & Related papers (2023-03-25T16:00:33Z) - Spatially-Adaptive Feature Modulation for Efficient Image
Super-Resolution [90.16462805389943]
We develop a spatially-adaptive feature modulation (SAFM) mechanism upon a vision transformer (ViT)-like block.
Proposed method is $3times$ smaller than state-of-the-art efficient SR methods.
arXiv Detail & Related papers (2023-02-27T14:19:31Z) - Self Sparse Generative Adversarial Networks [73.590634413751]
Generative Adversarial Networks (GANs) are an unsupervised generative model that learns data distribution through adversarial training.
We propose a Self Sparse Generative Adversarial Network (Self-Sparse GAN) that reduces the parameter space and alleviates the zero gradient problem.
arXiv Detail & Related papers (2021-01-26T04:49:12Z) - Lightweight Single-Image Super-Resolution Network with Attentive
Auxiliary Feature Learning [73.75457731689858]
We develop a computation efficient yet accurate network based on the proposed attentive auxiliary features (A$2$F) for SISR.
Experimental results on large-scale dataset demonstrate the effectiveness of the proposed model against the state-of-the-art (SOTA) SR methods.
arXiv Detail & Related papers (2020-11-13T06:01:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.