L3C-Stereo: Lossless Compression for Stereo Images
- URL: http://arxiv.org/abs/2108.09422v1
- Date: Sat, 21 Aug 2021 02:36:15 GMT
- Title: L3C-Stereo: Lossless Compression for Stereo Images
- Authors: Zihao Huang, Zhe Sun, Feng Duan, Andrzej Cichocki, Peiying Ruan and
Chao Li
- Abstract summary: A large number of autonomous driving tasks need high-definition stereo images, which requires a large amount of storage space.
To tackle this, we propose L3C-Stereo, a multi-scale lossless compression model consisting of two main modules.
In the experiments, our method outperforms the hand-crafted compression methods and the learning-based method on all three datasets used.
- Score: 24.6995551743392
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A large number of autonomous driving tasks need high-definition stereo
images, which requires a large amount of storage space. Efficiently executing
lossless compression has become a practical problem. Commonly, it is hard to
make accurate probability estimates for each pixel. To tackle this, we propose
L3C-Stereo, a multi-scale lossless compression model consisting of two main
modules: the warping module and the probability estimation module. The warping
module takes advantage of two view feature maps from the same domain to
generate a disparity map, which is used to reconstruct the right view so as to
improve the confidence of the probability estimate of the right view. The
probability estimation module provides pixel-wise logistic mixture
distributions for adaptive arithmetic coding. In the experiments, our method
outperforms the hand-crafted compression methods and the learning-based method
on all three datasets used. Then, we show that a better maximum disparity can
lead to a better compression effect. Furthermore, thanks to a compression
property of our model, it naturally generates a disparity map of an acceptable
quality for the subsequent stereo tasks.
Related papers
- PCF-Lift: Panoptic Lifting by Probabilistic Contrastive Fusion [80.79938369319152]
We design a new pipeline coined PCF-Lift based on our Probabilis-tic Contrastive Fusion (PCF)
Our PCF-lift not only significantly outperforms the state-of-the-art methods on widely used benchmarks including the ScanNet dataset and the Messy Room dataset (4.4% improvement of scene-level PQ)
arXiv Detail & Related papers (2024-10-14T16:06:59Z) - Variable-Rate Learned Image Compression with Multi-Objective
Optimization and Quantization-Reconstruction Offsets [8.670873561640903]
This paper follows the traditional approach to vary a single quantization step size to perform uniform quantization of all latent tensor elements.
Three modifications are proposed to improve the variable rate compression performance.
The achieved variable rate compression results indicate negligible or minimal compression performance loss compared to training multiple models.
arXiv Detail & Related papers (2024-02-29T07:45:02Z) - Generative Multiplane Neural Radiance for 3D-Aware Image Generation [102.15322193381617]
We present a method to efficiently generate 3D-aware high-resolution images that are view-consistent across multiple target views.
Our GMNR model generates 3D-aware images of 1024 X 1024 pixels with 17.6 FPS on a single V100.
arXiv Detail & Related papers (2023-04-03T17:41:20Z) - Selective compression learning of latent representations for
variable-rate image compression [38.077284943341105]
We propose a selective compression method that partially encodes latent representations in a fully generalized manner for deep learning-based variable-rate image compression.
The proposed method can achieve comparable compression efficiency as those of the separately trained reference compression models and can reduce decoding time owing to the selective compression.
arXiv Detail & Related papers (2022-11-08T09:09:59Z) - Curvature-guided dynamic scale networks for Multi-view Stereo [10.667165962654996]
This paper focuses on learning a robust feature extraction network to enhance the performance of matching costs without heavy computation.
We present a dynamic scale feature extraction network, namely, CDSFNet.
It is composed of multiple novel convolution layers, each of which can select a proper patch scale for each pixel guided by the normal curvature of the image surface.
arXiv Detail & Related papers (2021-12-11T14:41:05Z) - Variable-Rate Deep Image Compression through Spatially-Adaptive Feature
Transform [58.60004238261117]
We propose a versatile deep image compression network based on Spatial Feature Transform (SFT arXiv:1804.02815)
Our model covers a wide range of compression rates using a single model, which is controlled by arbitrary pixel-wise quality maps.
The proposed framework allows us to perform task-aware image compressions for various tasks.
arXiv Detail & Related papers (2021-08-21T17:30:06Z) - Learning Scalable $\ell_\infty$-constrained Near-lossless Image
Compression via Joint Lossy Image and Residual Compression [118.89112502350177]
We propose a novel framework for learning $ell_infty$-constrained near-lossless image compression.
We derive the probability model of the quantized residual by quantizing the learned probability model of the original residual.
arXiv Detail & Related papers (2021-03-31T11:53:36Z) - Displacement-Invariant Cost Computation for Efficient Stereo Matching [122.94051630000934]
Deep learning methods have dominated stereo matching leaderboards by yielding unprecedented disparity accuracy.
But their inference time is typically slow, on the order of seconds for a pair of 540p images.
We propose a emphdisplacement-invariant cost module to compute the matching costs without needing a 4D feature volume.
arXiv Detail & Related papers (2020-12-01T23:58:16Z) - Training with Quantization Noise for Extreme Model Compression [57.51832088938618]
We tackle the problem of producing compact models, maximizing their accuracy for a given model size.
A standard solution is to train networks with Quantization Aware Training, where the weights are quantized during training and the gradients approximated with the Straight-Through Estimator.
In this paper, we extend this approach to work beyond int8 fixed-point quantization with extreme compression methods.
arXiv Detail & Related papers (2020-04-15T20:10:53Z) - Compressed Volumetric Heatmaps for Multi-Person 3D Pose Estimation [33.71628590745982]
We present a novel approach for bottom-up multi-person 3D human pose estimation from monocular RGB images.
We propose a simple and effective compression method to drastically reduce the size of this representation.
Our method performs favorably when compared to state of the art on both multi-person and single-person 3D human pose estimation datasets.
arXiv Detail & Related papers (2020-04-01T10:37:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.