Neural Network Assisted Lifting Steps For Improved Fully Scalable Lossy
Image Compression in JPEG 2000
- URL: http://arxiv.org/abs/2403.01647v1
- Date: Mon, 4 Mar 2024 00:01:52 GMT
- Title: Neural Network Assisted Lifting Steps For Improved Fully Scalable Lossy
Image Compression in JPEG 2000
- Authors: Xinyue Li, Aous Naman and David Taubman
- Abstract summary: This work proposes to augment the lifting steps of the conventional wavelet transform with additional neural network assisted lifting steps.
The proposed approach involves two steps, a high-to-low step followed by a low-to-high step.
By employing the proposed approach within the JPEG 2000 image coding standard, our method can achieve up to 17.4% average BD bit-rate saving.
- Score: 14.473452842448737
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This work proposes to augment the lifting steps of the conventional wavelet
transform with additional neural network assisted lifting steps. These
additional steps reduce residual redundancy (notably aliasing information)
amongst the wavelet subbands, and also improve the visual quality of
reconstructed images at reduced resolutions. The proposed approach involves two
steps, a high-to-low step followed by a low-to-high step. The high-to-low step
suppresses aliasing in the low-pass band by using the detail bands at the same
resolution, while the low-to-high step aims to further remove redundancy from
detail bands, so as to achieve higher energy compaction. The proposed two
lifting steps are trained in an end-to-end fashion; we employ a backward
annealing approach to overcome the non-differentiability of the quantization
and cost functions during back-propagation. Importantly, the networks employed
in this paper are compact and with limited non-linearities, allowing a fully
scalable system; one pair of trained network parameters are applied for all
levels of decomposition and for all bit-rates of interest. By employing the
proposed approach within the JPEG 2000 image coding standard, our method can
achieve up to 17.4% average BD bit-rate saving over a wide range of bit-rates,
while retaining quality and resolution scalability features of JPEG 2000.
Related papers
- PassionSR: Post-Training Quantization with Adaptive Scale in One-Step Diffusion based Image Super-Resolution [87.89013794655207]
Diffusion-based image super-resolution (SR) models have shown superior performance at the cost of multiple denoising steps.
We propose a novel post-training quantization approach with adaptive scale in one-step diffusion (OSD) image SR, PassionSR.
Our PassionSR achieves significant advantages over recent leading low-bit quantization methods for image SR.
arXiv Detail & Related papers (2024-11-26T04:49:42Z) - Learning Efficient and Effective Trajectories for Differential Equation-based Image Restoration [59.744840744491945]
We reformulate the trajectory optimization of this kind of method, focusing on enhancing both reconstruction quality and efficiency.
We propose cost-aware trajectory distillation to streamline complex paths into several manageable steps with adaptable sizes.
Experiments showcase the significant superiority of the proposed method, achieving a maximum PSNR improvement of 2.1 dB over state-of-the-art methods.
arXiv Detail & Related papers (2024-10-07T07:46:08Z) - 2DQuant: Low-bit Post-Training Quantization for Image Super-Resolution [83.09117439860607]
Low-bit quantization has become widespread for compressing image super-resolution (SR) models for edge deployment.
It is notorious that low-bit quantization degrades the accuracy of SR models compared to their full-precision (FP) counterparts.
We present a dual-stage low-bit post-training quantization (PTQ) method for image super-resolution, namely 2DQuant, which achieves efficient and accurate SR under low-bit quantization.
arXiv Detail & Related papers (2024-06-10T06:06:11Z) - Frequency-Aware Re-Parameterization for Over-Fitting Based Image
Compression [12.725194101094711]
Over-fitting-based image compression requires weights compactness for compression and fast convergence for practical use.
This paper presents a simple re- parameterization method to train CNNs with reduced weights storage and accelerated convergence.
The proposed method is verified with extensive experiments of over-fitting-based image restoration on various datasets, achieving up to -46.12% BD-rate on top of HEIF with only 200 iterations.
arXiv Detail & Related papers (2023-10-12T06:32:12Z) - Learned Image Compression with Generalized Octave Convolution and
Cross-Resolution Parameter Estimation [5.238765582868391]
We propose a learned multi-resolution image compression framework, which exploits octave convolutions to factorize the latent representations into the high-resolution (HR) and low-resolution (LR) parts.
Experimental results show that our method separately reduces the decoding time by approximately 73.35 % and 93.44 % compared with that of state-of-the-art learned image compression methods.
arXiv Detail & Related papers (2022-09-07T08:21:52Z) - Dynamic Dual Trainable Bounds for Ultra-low Precision Super-Resolution
Networks [82.18396309806577]
We propose a novel activation quantizer, referred to as Dynamic Dual Trainable Bounds (DDTB)
Our DDTB exhibits significant performance improvements in ultra-low precision.
For example, our DDTB achieves a 0.70dB PSNR increase on Urban100 benchmark when quantizing EDSR to 2-bit and scaling up output images to x4.
arXiv Detail & Related papers (2022-03-08T04:26:18Z) - Invertible Network for Unpaired Low-light Image Enhancement [78.33382003460903]
We propose to leverage the invertible network to enhance low-light image in forward process and degrade the normal-light one inversely with unpaired learning.
In addition to the adversarial loss, we design various loss functions to ensure the stability of training and preserve more image details.
We present a progressive self-guided enhancement process for low-light images and achieve favorable performance against the SOTAs.
arXiv Detail & Related papers (2021-12-24T17:00:54Z) - Efficient Re-parameterization Residual Attention Network For
Nonhomogeneous Image Dehazing [4.723586858098229]
ERRA-Net has an impressive speed, processing 1200x1600 HD quality images with an average runtime of 166.11 fps.
We use cascaded MA blocks to extract high-frequency features step by step, and the Multi-layer attention fusion tail combines the shallow and deep features of the model to get the residual of the clean image.
arXiv Detail & Related papers (2021-09-12T10:03:44Z) - Remote Sensing Image Scene Classification with Deep Neural Networks in
JPEG 2000 Compressed Domain [8.296684637620553]
Existing scene classification approaches using deep neural networks (DNNs) require to fully decompress the images.
We propose a novel approach to achieve scene classification in JPEG 2000 compressed RS images.
arXiv Detail & Related papers (2020-06-20T09:13:38Z) - Hyperspectral Image Super-resolution via Deep Progressive Zero-centric
Residual Learning [62.52242684874278]
Cross-modality distribution of spatial and spectral information makes the problem challenging.
We propose a novel textitlightweight deep neural network-based framework, namely PZRes-Net.
Our framework learns a high resolution and textitzero-centric residual image, which contains high-frequency spatial details of the scene.
arXiv Detail & Related papers (2020-06-18T06:32:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.