A Multiparametric Class of Low-complexity Transforms for Image and Video
Coding
- URL: http://arxiv.org/abs/2006.11418v1
- Date: Fri, 19 Jun 2020 21:56:58 GMT
- Title: A Multiparametric Class of Low-complexity Transforms for Image and Video
Coding
- Authors: D. R. Canterle, T. L. T. da Silveira, F. M. Bayer, R. J. Cintra
- Abstract summary: We introduce a new class of low-complexity 8-point DCT approximations based on a series of works published by Bouguezel, Ahmed and Swamy.
We show that the optimal DCT approximations present compelling results in terms of coding efficiency and image quality metrics.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Discrete transforms play an important role in many signal processing
applications, and low-complexity alternatives for classical transforms became
popular in recent years. Particularly, the discrete cosine transform (DCT) has
proven to be convenient for data compression, being employed in well-known
image and video coding standards such as JPEG, H.264, and the recent high
efficiency video coding (HEVC). In this paper, we introduce a new class of
low-complexity 8-point DCT approximations based on a series of works published
by Bouguezel, Ahmed and Swamy. Also, a multiparametric fast algorithm that
encompasses both known and novel transforms is derived. We select the
best-performing DCT approximations after solving a multicriteria optimization
problem, and submit them to a scaling method for obtaining larger size
transforms. We assess these DCT approximations in both JPEG-like image
compression and video coding experiments. We show that the optimal DCT
approximations present compelling results in terms of coding efficiency and
image quality metrics, and require only few addition or bit-shifting
operations, being suitable for low-complexity and low-power systems.
Related papers
- Accelerating Error Correction Code Transformers [56.75773430667148]
We introduce a novel acceleration method for transformer-based decoders.
We achieve a 90% compression ratio and reduce arithmetic operation energy consumption by at least 224 times on modern hardware.
arXiv Detail & Related papers (2024-10-08T11:07:55Z) - Progressive Learning with Visual Prompt Tuning for Variable-Rate Image
Compression [60.689646881479064]
We propose a progressive learning paradigm for transformer-based variable-rate image compression.
Inspired by visual prompt tuning, we use LPM to extract prompts for input images and hidden features at the encoder side and decoder side, respectively.
Our model outperforms all current variable image methods in terms of rate-distortion performance and approaches the state-of-the-art fixed image compression methods trained from scratch.
arXiv Detail & Related papers (2023-11-23T08:29:32Z) - AICT: An Adaptive Image Compression Transformer [18.05997169440533]
We propose a more straightforward yet effective Tranformer-based channel-wise auto-regressive prior model, resulting in an absolute image compression transformer (ICT)
The proposed ICT can capture both global and local contexts from the latent representations.
We leverage a learnable scaling module with a sandwich ConvNeXt-based pre/post-processor to accurately extract more compact latent representation.
arXiv Detail & Related papers (2023-07-12T11:32:02Z) - Low-Complexity Loeffler DCT Approximations for Image and Video Coding [0.0]
This paper introduces a matrix parametrization method based on the Loeffler discrete cosine transform (DCT) algorithm.
A new class of eight-point DCT approximations was proposed, capable of unifying the mathematical formalism of several eight-point DCT approximations.
arXiv Detail & Related papers (2022-07-29T03:56:18Z) - DCT Approximations Based on Chen's Factorization [0.17205106391379021]
Two 8-point multiplication-free DCT approximations are proposed and their fast algorithms are also derived.
Experiments with a JPEG-like image compression scheme are performed and results are compared with competing methods.
New sets of transformations are embedded into an HEVC reference software to provide a fully HEVC-compliant video coding scheme.
arXiv Detail & Related papers (2022-07-24T02:31:28Z) - Neural Data-Dependent Transform for Learned Image Compression [72.86505042102155]
We build a neural data-dependent transform and introduce a continuous online mode decision mechanism to jointly optimize the coding efficiency for each individual image.
The experimental results show the effectiveness of the proposed neural-syntax design and the continuous online mode decision mechanism.
arXiv Detail & Related papers (2022-03-09T14:56:48Z) - Neural JPEG: End-to-End Image Compression Leveraging a Standard JPEG
Encoder-Decoder [73.48927855855219]
We propose a system that learns to improve the encoding performance by enhancing its internal neural representations on both the encoder and decoder ends.
Experiments demonstrate that our approach successfully improves the rate-distortion performance over JPEG across various quality metrics.
arXiv Detail & Related papers (2022-01-27T20:20:03Z) - Towards End-to-End Image Compression and Analysis with Transformers [99.50111380056043]
We propose an end-to-end image compression and analysis model with Transformers, targeting to the cloud-based image classification application.
We aim to redesign the Vision Transformer (ViT) model to perform image classification from the compressed features and facilitate image compression with the long-term information from the Transformer.
Experimental results demonstrate the effectiveness of the proposed model in both the image compression and the classification tasks.
arXiv Detail & Related papers (2021-12-17T03:28:14Z) - Generating Images with Sparse Representations [21.27273495926409]
High dimensionality of images presents architecture and sampling-efficiency challenges for likelihood-based generative models.
We present an alternative approach, inspired by common image compression methods like JPEG, and convert images to quantized discrete cosine transform (DCT) blocks.
We propose a Transformer-based autoregressive architecture, which is trained to sequentially predict the conditional distribution of the next element in such sequences.
arXiv Detail & Related papers (2021-03-05T17:56:03Z) - Learned Multi-Resolution Variable-Rate Image Compression with
Octave-based Residual Blocks [15.308823742699039]
We propose a new variable-rate image compression framework, which employs generalized octave convolutions (GoConv) and generalized octave transposed-convolutions (GoTConv)
To enable a single model to operate with different bit rates and to learn multi-rate image features, a new objective function is introduced.
Experimental results show that the proposed framework trained with variable-rate objective function outperforms the standard codecs such as H.265/HEVC-based BPG and state-of-the-art learning-based variable-rate methods.
arXiv Detail & Related papers (2020-12-31T06:26:56Z) - Learning to Improve Image Compression without Changing the Standard
Decoder [100.32492297717056]
We propose learning to improve the encoding performance with the standard decoder.
Specifically, a frequency-domain pre-editing method is proposed to optimize the distribution of DCT coefficients.
We do not modify the JPEG decoder and therefore our approach is applicable when viewing images with the widely used standard JPEG decoder.
arXiv Detail & Related papers (2020-09-27T19:24:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.