Light Field Compression by Residual CNN Assisted JPEG
- URL: http://arxiv.org/abs/2010.00062v2
- Date: Thu, 18 Mar 2021 15:02:06 GMT
- Title: Light Field Compression by Residual CNN Assisted JPEG
- Authors: Eisa Hedayati, Timothy C. Havens, Jeremy P. Bos
- Abstract summary: We develop a JPEG-assisted learning-based technique to reconstruct an LF from a bitstream with a bit per pixel ratio of 0.0047 on average.
We show that with a 1% compression time cost and 18x speedup for decompression, our methods reconstructed LFs have better structural similarity index metric (SSIM) and comparable peak signal-to-noise ratio (PSNR) compared to the state-of-the-art video compression techniques used to compress LFs.
- Score: 4.767599257804181
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Light field (LF) imaging has gained significant attention due to its recent
success in 3-dimensional (3D) displaying and rendering as well as augmented and
virtual reality usage. Nonetheless, because of the two extra dimensions, LFs
are much larger than conventional images. We develop a JPEG-assisted
learning-based technique to reconstruct an LF from a JPEG bitstream with a bit
per pixel ratio of 0.0047 on average. For compression, we keep the LF's center
view and use JPEG compression with 50% quality. Our reconstruction pipeline
consists of a small JPEG enhancement network (JPEG-Hance), a depth estimation
network (Depth-Net), followed by view synthesizing by warping the enhanced
center view. Our pipeline is significantly faster than using video compression
on pseudo-sequences extracted from an LF, both in compression and
decompression, while maintaining effective performance. We show that with a 1%
compression time cost and 18x speedup for decompression, our methods
reconstructed LFs have better structural similarity index metric (SSIM) and
comparable peak signal-to-noise ratio (PSNR) compared to the state-of-the-art
video compression techniques used to compress LFs.
Related papers
- LFIC-DRASC: Deep Light Field Image Compression Using Disentangled Representation and Asymmetrical Strip Convolution [51.909036244222904]
We propose an end-to-end deep LF Image Compression method using Disentangled Representation and Asymmetrical Strip Convolution.
Experimental results demonstrate that the proposed LFIC-DRASC achieves an average of 20.5% bit rate reductions.
arXiv Detail & Related papers (2024-09-18T05:33:42Z) - Learned Image Compression for HE-stained Histopathological Images via Stain Deconvolution [33.69980388844034]
In this paper, we show that the commonly used JPEG algorithm is not best suited for further compression.
We propose Stain Quantized Latent Compression, a novel DL based histopathology data compression approach.
We show that our approach yields superior performance in a classification downstream task, compared to traditional approaches like JPEG.
arXiv Detail & Related papers (2024-06-18T13:47:17Z) - RAGE for the Machine: Image Compression with Low-Cost Random Access for
Embedded Applications [5.199703527082964]
RAGE is an image compression framework that achieves four generally conflicting objectives.
We show that RAGE has similar or better compression ratios to state-of-the-art lossless image compressors.
Our measurements also show that RAGE's lossy variant, RAGE-Q, outperforms JPEG by several fold in terms of distortion in embedded graphics.
arXiv Detail & Related papers (2024-02-07T19:28:33Z) - Unified learning-based lossy and lossless JPEG recompression [15.922937139019547]
We propose a unified lossly and lossless JPEG recompression framework, which consists of learned quantization table and Markovian hierarchical variational autoencoders.
Experiments show that our method can achieve arbitrarily low distortion when the JPEG is close to the upper bound.
arXiv Detail & Related papers (2023-12-05T12:07:27Z) - Learned Lossless Compression for JPEG via Frequency-Domain Prediction [50.20577108662153]
We propose a novel framework for learned lossless compression of JPEG images.
To enable learning in the frequency domain, DCT coefficients are partitioned into groups to utilize implicit local redundancy.
An autoencoder-like architecture is designed based on the weight-shared blocks to realize entropy modeling of grouped DCT coefficients.
arXiv Detail & Related papers (2023-03-05T13:15:28Z) - Data Efficient Visual Place Recognition Using Extremely JPEG-Compressed
Images [17.847661026367767]
This paper studies the effects of JPEG compression on the performance of Visual Place Recognition techniques.
We show that by introducing compression, the VPR performance is drastically reduced, especially in the higher spectrum of compression.
We present a fine-tuned CNN which is optimized for JPEG compressed data and show that it performs more consistently with the image transformations detected in extremely compressed JPEG images.
arXiv Detail & Related papers (2022-09-17T14:46:28Z) - Neural JPEG: End-to-End Image Compression Leveraging a Standard JPEG
Encoder-Decoder [73.48927855855219]
We propose a system that learns to improve the encoding performance by enhancing its internal neural representations on both the encoder and decoder ends.
Experiments demonstrate that our approach successfully improves the rate-distortion performance over JPEG across various quality metrics.
arXiv Detail & Related papers (2022-01-27T20:20:03Z) - Learning to Improve Image Compression without Changing the Standard
Decoder [100.32492297717056]
We propose learning to improve the encoding performance with the standard decoder.
Specifically, a frequency-domain pre-editing method is proposed to optimize the distribution of DCT coefficients.
We do not modify the JPEG decoder and therefore our approach is applicable when viewing images with the widely used standard JPEG decoder.
arXiv Detail & Related papers (2020-09-27T19:24:42Z) - Quantization Guided JPEG Artifact Correction [69.04777875711646]
We develop a novel architecture for artifact correction using the JPEG files quantization matrix.
This allows our single model to achieve state-of-the-art performance over models trained for specific quality settings.
arXiv Detail & Related papers (2020-04-17T00:10:08Z) - Learning Better Lossless Compression Using Lossy Compression [100.50156325096611]
We leverage the powerful lossy image compression algorithm BPG to build a lossless image compression system.
We model the distribution of the residual with a convolutional neural network-based probabilistic model that is conditioned on the BPG reconstruction.
Finally, the image is stored using the concatenation of the bitstreams produced by BPG and the learned residual coder.
arXiv Detail & Related papers (2020-03-23T11:21:52Z) - Improved Image Coding Autoencoder With Deep Learning [8.92071749364712]
We build autoencoder based pipelines for extreme end-to-end image compression based on Ball'e's approach.
It achieved around 4.0% reduction in bits per pixel (bpp), 0.03% increase in multi-scale structural similarity (MS-SSIM) and only 0.47% decrease in peak signal-to-noise ratio (PSNR)
arXiv Detail & Related papers (2020-02-28T03:21:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.