Point Cloud Geometry Scalable Coding with a Quality-Conditioned Latents Probability Estimator
- URL: http://arxiv.org/abs/2404.07698v2
- Date: Tue, 9 Jul 2024 06:56:06 GMT
- Title: Point Cloud Geometry Scalable Coding with a Quality-Conditioned Latents Probability Estimator
- Authors: Daniele Mari, André F. R. Guarda, Nuno M. M. Rodrigues, Simone Milani, Fernando Pereira,
- Abstract summary: Quality scalability is a major requirement in most learning-based PC coding solutions.
This paper proposes a quality scalability scheme, named Scalable Quality Hyperprior (SQH), adaptable to learning-based static point cloud geometry codecs.
SQH offers the quality scalability feature with very limited or no compression performance penalty at all when compared with the corresponding non-scalable solution.
- Score: 47.792286013837945
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The widespread usage of point clouds (PC) for immersive visual applications has resulted in the use of very heterogeneous receiving conditions and devices, notably in terms of network, hardware, and display capabilities. In this scenario, quality scalability, i.e., the ability to reconstruct a signal at different qualities by progressively decoding a single bitstream, is a major requirement that has yet to be conveniently addressed, notably in most learning-based PC coding solutions. This paper proposes a quality scalability scheme, named Scalable Quality Hyperprior (SQH), adaptable to learning-based static point cloud geometry codecs, which uses a Quality-conditioned Latents Probability Estimator (QuLPE) to decode a high-quality version of a PC learning-based representation, based on an available lower quality base layer. SQH is integrated in the future JPEG PC coding standard, allowing to create a layered bitstream that can be used to progressively decode the PC geometry with increasing quality and fidelity. Experimental results show that SQH offers the quality scalability feature with very limited or no compression performance penalty at all when compared with the corresponding non-scalable solution, thus preserving the significant compression gains over other state-of-the-art PC codecs.
Related papers
- Point Cloud Geometry Scalable Coding Using a Resolution and Quality-conditioned Latents Probability Estimator [47.792286013837945]
This paper focuses on the development of scalable coding solutions for deep learning-based Point Cloud (PC) coding.
The peculiarities of this 3D representation make it hard to implement flexible solutions that do not compromise the other functionalities of the software.
arXiv Detail & Related papers (2025-02-19T20:58:53Z) - Prediction and Reference Quality Adaptation for Learned Video Compression [54.58691829087094]
Temporal prediction is one of the most important technologies for video compression.
Traditional video codecs adaptively decide the optimal coding mode according to the prediction quality and reference quality.
We propose a confidence-based prediction quality adaptation (PQA) module and a reference quality adaptation (RQA) module.
arXiv Detail & Related papers (2024-06-20T09:03:26Z) - Contrastive Pre-Training with Multi-View Fusion for No-Reference Point Cloud Quality Assessment [49.36799270585947]
No-reference point cloud quality assessment (NR-PCQA) aims to automatically evaluate the perceptual quality of distorted point clouds without available reference.
We propose a novel contrastive pre-training framework tailored for PCQA (CoPA)
Our method outperforms the state-of-the-art PCQA methods on popular benchmarks.
arXiv Detail & Related papers (2024-03-15T07:16:07Z) - Simple Baselines for Projection-based Full-reference and No-reference
Point Cloud Quality Assessment [60.2709006613171]
We propose simple baselines for projection-based point cloud quality assessment (PCQA)
We use multi-projections obtained via a common cube-like projection process from the point clouds for both full-reference (FR) and no-reference (NR) PCQA tasks.
Taking part in the ICIP 2023 PCVQA Challenge, we succeeded in achieving the top spot in four out of the five competition tracks.
arXiv Detail & Related papers (2023-10-26T04:42:57Z) - Extreme Image Compression using Fine-tuned VQGANs [43.43014096929809]
We introduce vector quantization (VQ)-based generative models into the image compression domain.
The codebook learned by the VQGAN model yields a strong expressive capacity.
The proposed framework outperforms state-of-the-art codecs in terms of perceptual quality-oriented metrics.
arXiv Detail & Related papers (2023-07-17T06:14:19Z) - Deep probabilistic model for lossless scalable point cloud attribute
compression [2.2559617939136505]
We build an end-to-end point cloud attribute coding method (MNeT) that progressively projects the attributes onto multiscale latent spaces.
We validate our method on a set of point clouds from MVUB and MPEG and show that our method outperforms recently proposed methods and on par with the latest G-PCC version 14.
arXiv Detail & Related papers (2023-03-11T23:39:30Z) - Reduced Reference Perceptual Quality Model and Application to Rate
Control for 3D Point Cloud Compression [61.110938359555895]
In rate-distortion optimization, the encoder settings are determined by maximizing a reconstruction quality measure subject to a constraint on the bit rate.
We propose a linear perceptual quality model whose variables are the V-PCC geometry and color quantization parameters.
Subjective quality tests with 400 compressed 3D point clouds show that the proposed model correlates well with the mean opinion score.
We show that for the same target bit rate, ratedistortion optimization based on the proposed model offers higher perceptual quality than rate-distortion optimization based on exhaustive search with a point-to-point objective quality metric.
arXiv Detail & Related papers (2020-11-25T12:42:02Z) - Multiscale Point Cloud Geometry Compression [29.605320327889142]
We propose a multiscale-to-end learning framework which hierarchically reconstructs the 3D Point Cloud Geometry.
The framework is developed on top of a sparse convolution based autoencoder for point cloud compression and reconstruction.
arXiv Detail & Related papers (2020-11-07T16:11:16Z) - Adapting JPEG XS gains and priorities to tasks and contents [34.99683302788977]
Constant market requirements for a low-complexity image have led to the recent development and standardization of a lightweight image named JPEG XS.
In this work we show that JPEG XS compression can be adapted to a given task and content, such as preserving visual quality on desktop content or maintaining high accuracy in neural network segmentation tasks.
arXiv Detail & Related papers (2020-05-18T14:33:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.