Inter-Frame Compression for Dynamic Point Cloud Geometry Coding
- URL: http://arxiv.org/abs/2207.12554v1
- Date: Mon, 25 Jul 2022 22:17:19 GMT
- Title: Inter-Frame Compression for Dynamic Point Cloud Geometry Coding
- Authors: Anique Akhtar, Zhu Li, Geert Van der Auwera
- Abstract summary: We propose a lossy compression scheme that predicts the latent representation of the current frame using the previous frame.
Our method achieves more than 91% BD-Rate Bjontegaard Delta Rate and more than 62% BD-Rate reduction against V-PCC intra-frame encoding mode.
- Score: 9.15965133212928
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Efficient point cloud compression is essential for applications like virtual
and mixed reality, autonomous driving, and cultural heritage. In this paper, we
propose a deep learning-based inter-frame encoding scheme for dynamic point
cloud geometry compression. We propose a lossy geometry compression scheme that
predicts the latent representation of the current frame using the previous
frame by employing a novel prediction network. Our proposed network utilizes
sparse convolutions with hierarchical multiscale 3D feature learning to encode
the current frame using the previous frame. We employ convolution on target
coordinates to map the latent representation of the previous frame to the
downsampled coordinates of the current frame to predict the current frame's
feature embedding. Our framework transmits the residual of the predicted
features and the actual features by compressing them using a learned
probabilistic factorized entropy model. At the receiver, the decoder
hierarchically reconstructs the current frame by progressively rescaling the
feature embedding. We compared our model to the state-of-the-art Video-based
Point Cloud Compression (V-PCC) and Geometry-based Point Cloud Compression
(G-PCC) schemes standardized by the Moving Picture Experts Group (MPEG). Our
method achieves more than 91% BD-Rate Bjontegaard Delta Rate) reduction against
G-PCC, more than 62% BD-Rate reduction against V-PCC intra-frame encoding mode,
and more than 52% BD-Rate savings against V-PCC P-frame-based inter-frame
encoding mode using HEVC.
Related papers
- Hierarchical Prior-based Super Resolution for Point Cloud Geometry
Compression [39.052583172727324]
The Geometry-based Point Cloud Compression (G-PCC) has been developed by the Moving Picture Experts Group to compress point clouds.
This paper proposes a hierarchical prior-based super resolution method for point cloud geometry compression.
arXiv Detail & Related papers (2024-02-17T11:15:38Z) - IBVC: Interpolation-driven B-frame Video Compression [68.18440522300536]
B-frame video compression aims to adopt bi-directional motion estimation and motion compensation (MEMC) coding for middle frame reconstruction.
Previous learned approaches often directly extend neural P-frame codecs to B-frame relying on bi-directional optical-flow estimation.
We propose a simple yet effective structure called Interpolation-B-frame Video Compression (IBVC) to address these issues.
arXiv Detail & Related papers (2023-09-25T02:45:51Z) - VNVC: A Versatile Neural Video Coding Framework for Efficient
Human-Machine Vision [59.632286735304156]
It is more efficient to enhance/analyze the coded representations directly without decoding them into pixels.
We propose a versatile neural video coding (VNVC) framework, which targets learning compact representations to support both reconstruction and direct enhancement/analysis.
arXiv Detail & Related papers (2023-06-19T03:04:57Z) - Dynamic Point Cloud Geometry Compression Using Multiscale Inter
Conditional Coding [27.013814232906817]
This work extends the Multiscale Sparse Representation (MSR) framework developed for Point Cloud Geometry Compression (PCGC) to support the dynamic PCGC.
The reconstruction of the preceding Point Cloud Geometry (PCG) frame is progressively downscaled to generate multiscale temporal priors.
arXiv Detail & Related papers (2023-01-28T11:34:06Z) - FFNeRV: Flow-Guided Frame-Wise Neural Representations for Videos [5.958701846880935]
We propose FFNeRV, a novel method for incorporating flow information into frame-wise representations to exploit the temporal redundancy across the frames in videos.
With model compression techniques, FFNeRV outperforms widely-used standard video codecs (H.264 and HEVC) and performs on par with state-of-the-art video compression algorithms.
arXiv Detail & Related papers (2022-12-23T12:51:42Z) - Learned Video Compression via Heterogeneous Deformable Compensation
Network [78.72508633457392]
We propose a learned video compression framework via heterogeneous deformable compensation strategy (HDCVC) to tackle the problems of unstable compression performance.
More specifically, the proposed algorithm extracts features from the two adjacent frames to estimate content-Neighborhood heterogeneous deformable (HetDeform) kernel offsets.
Experimental results indicate that HDCVC achieves superior performance than the recent state-of-the-art learned video compression approaches.
arXiv Detail & Related papers (2022-07-11T02:31:31Z) - Point Cloud Compression with Sibling Context and Surface Priors [47.96018990521301]
We present a novel octree-based multi-level framework for large-scale point cloud compression.
In this framework, we propose a new entropy model that explores the hierarchical dependency in an octree.
We locally fit surfaces with a voxel-based geometry-aware module to provide geometric priors in entropy encoding.
arXiv Detail & Related papers (2022-05-02T09:13:26Z) - Multiscale Point Cloud Geometry Compression [29.605320327889142]
We propose a multiscale-to-end learning framework which hierarchically reconstructs the 3D Point Cloud Geometry.
The framework is developed on top of a sparse convolution based autoencoder for point cloud compression and reconstruction.
arXiv Detail & Related papers (2020-11-07T16:11:16Z) - Conditional Entropy Coding for Efficient Video Compression [82.35389813794372]
We propose a very simple and efficient video compression framework that only focuses on modeling the conditional entropy between frames.
We first show that a simple architecture modeling the entropy between the image latent codes is as competitive as other neural video compression works and video codecs.
We then propose a novel internal learning extension on top of this architecture that brings an additional 10% savings without trading off decoding speed.
arXiv Detail & Related papers (2020-08-20T20:01:59Z) - End-to-End Learning for Video Frame Compression with Self-Attention [25.23586503813838]
We propose an end-to-end learned system for compressing video frames.
Our system learns deep embeddings of frames and encodes their difference in latent space.
In our experiments, we show that the proposed system achieves high compression rates and high objective visual quality.
arXiv Detail & Related papers (2020-04-20T12:11:08Z) - Learning for Video Compression with Hierarchical Quality and Recurrent
Enhancement [164.7489982837475]
We propose a Hierarchical Learned Video Compression (HLVC) method with three hierarchical quality layers and a recurrent enhancement network.
In our HLVC approach, the hierarchical quality benefits the coding efficiency, since the high quality information facilitates the compression and enhancement of low quality frames at encoder and decoder sides.
arXiv Detail & Related papers (2020-03-04T09:31:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.