DPICT: Deep Progressive Image Compression Using Trit-Planes
- URL: http://arxiv.org/abs/2112.06334v1
- Date: Sun, 12 Dec 2021 22:09:33 GMT
- Title: DPICT: Deep Progressive Image Compression Using Trit-Planes
- Authors: Jae-Han Lee, Seungmin Jeon, Kwang Pyo Choi, Youngo Park, and Chang-Su
Kim
- Abstract summary: Deep progressive image compression using trit-planes (DPICT) algorithm.
We transform an image into a latent tensor using an analysis network.
We encode it into a compressed bitstream trit-plane by trit-plane in the decreasing order of significance.
- Score: 36.34865777731784
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose the deep progressive image compression using trit-planes (DPICT)
algorithm, which is the first learning-based codec supporting fine granular
scalability (FGS). First, we transform an image into a latent tensor using an
analysis network. Then, we represent the latent tensor in ternary digits
(trits) and encode it into a compressed bitstream trit-plane by trit-plane in
the decreasing order of significance. Moreover, within each trit-plane, we sort
the trits according to their rate-distortion priorities and transmit more
important information first. Since the compression network is less optimized
for the cases of using fewer trit-planes, we develop a postprocessing network
for refining reconstructed images at low rates. Experimental results show that
DPICT outperforms conventional progressive codecs significantly, while enabling
FGS transmission.
Related papers
- Chasing Better Deep Image Priors between Over- and Under-parameterization [63.8954152220162]
We study a novel "lottery image prior" (LIP) by exploiting DNN inherent sparsity.
LIPworks significantly outperform deep decoders under comparably compact model sizes.
We also extend LIP to compressive sensing image reconstruction, where a pre-trained GAN generator is used as the prior.
arXiv Detail & Related papers (2024-10-31T17:49:44Z) - Semantic Ensemble Loss and Latent Refinement for High-Fidelity Neural Image Compression [58.618625678054826]
This study presents an enhanced neural compression method designed for optimal visual fidelity.
We have trained our model with a sophisticated semantic ensemble loss, integrating Charbonnier loss, perceptual loss, style loss, and a non-binary adversarial loss.
Our empirical findings demonstrate that this approach significantly improves the statistical fidelity of neural image compression.
arXiv Detail & Related papers (2024-01-25T08:11:27Z) - Context-Based Trit-Plane Coding for Progressive Image Compression [31.396712329965005]
Trit-plane coding enables deep progressive image compression, but it cannot use autoregressive context models.
We develop the context-based rate reduction module to estimate trit probabilities of latent elements accurately.
Second, we develop the context-based distortion reduction module to refine partial latent tensors from the trit-planes.
Third, we propose a retraining scheme for the decoder to attain better rate-distortion tradeoffs.
arXiv Detail & Related papers (2023-03-10T05:46:25Z) - PREF: Phasorial Embedding Fields for Compact Neural Representations [54.44527545923917]
We present a phasorial embedding field emphPREF as a compact representation to facilitate neural signal modeling and reconstruction tasks.
Our experiments show PREF-based neural signal processing technique is on par with the state-of-the-art in 2D image completion, 3D SDF surface regression, and 5D radiance field reconstruction.
arXiv Detail & Related papers (2022-05-26T17:43:03Z) - RD-Optimized Trit-Plane Coding of Deep Compressed Image Latent Tensors [40.86513649546442]
DPICT is the first learning-based image supporting fine granular scalability.
In this paper, we describe how to implement two key components of DPICT efficiently: trit-plane slicing and RD-prioritized transmission.
arXiv Detail & Related papers (2022-03-25T06:33:16Z) - Reducing Redundancy in the Bottleneck Representation of the Autoencoders [98.78384185493624]
Autoencoders are a type of unsupervised neural networks, which can be used to solve various tasks.
We propose a scheme to explicitly penalize feature redundancies in the bottleneck representation.
We tested our approach across different tasks: dimensionality reduction using three different dataset, image compression using the MNIST dataset, and image denoising using fashion MNIST.
arXiv Detail & Related papers (2022-02-09T18:48:02Z) - Exploring Structural Sparsity in Neural Image Compression [14.106763725475469]
We propose a plug-in adaptive binary channel masking(ABCM) to judge the importance of each convolution channel and introduce sparsity during training.
During inference, the unimportant channels are pruned to obtain slimmer network and less computation.
Experiment results show that up to 7x computation reduction and 3x acceleration can be achieved with negligible performance drop.
arXiv Detail & Related papers (2022-02-09T17:46:49Z) - DeepFGS: Fine-Grained Scalable Coding for Learned Image Compression [22.933872281183497]
We propose the first learned fine-grained scalable image compression model (DeepFGS)
In this paper, we introduce a feature separation backbone to divide the image information into basic and scalable features, then redistribute the features channel by channel through an information rearrangement strategy.
Experiments demonstrate that our DeepFGS outperforms all learning-based scalable image compression models and conventional scalable image codecs in PSNR and MS-SSIM metrics.
arXiv Detail & Related papers (2022-01-04T15:03:13Z) - Modeling Image Quantization Tradeoffs for Optimal Compression [0.0]
Lossy compression algorithms target tradeoffs by quantizating high frequency data to increase compression rates.
We propose a new method of optimizing quantization tables using Deep Learning and a minimax loss function.
arXiv Detail & Related papers (2021-12-14T07:35:22Z) - Implicit Neural Representations for Image Compression [103.78615661013623]
Implicit Neural Representations (INRs) have gained attention as a novel and effective representation for various data types.
We propose the first comprehensive compression pipeline based on INRs including quantization, quantization-aware retraining and entropy coding.
We find that our approach to source compression with INRs vastly outperforms similar prior work.
arXiv Detail & Related papers (2021-12-08T13:02:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.