Lossless Compression of Point Cloud Sequences Using Sequence Optimized
CNN Models
- URL: http://arxiv.org/abs/2206.01297v1
- Date: Thu, 2 Jun 2022 20:46:05 GMT
- Title: Lossless Compression of Point Cloud Sequences Using Sequence Optimized
CNN Models
- Authors: Emre Can Kaya and Ioan Tabus
- Abstract summary: We propose a new paradigm for encoding the geometry of point cloud sequences, where the convolutional neural network estimates the encoding distributions.
We adopt lightweight CNN structures, we perform training as part of the encoding process, and the CNN parameters are transmitted as part of the bitstream.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We propose a new paradigm for encoding the geometry of point cloud sequences,
where the convolutional neural network (CNN) which estimates the encoding
distributions is optimized on several frames of the sequence to be compressed.
We adopt lightweight CNN structures, we perform training as part of the
encoding process, and the CNN parameters are transmitted as part of the
bitstream. The newly proposed encoding scheme operates on the octree
representation for each point cloud, encoding consecutively each octree
resolution layer. At every octree resolution layer, the voxel grid is traversed
section-by-section (each section being perpendicular to a selected coordinate
axis) and in each section the occupancies of groups of two-by-two voxels are
encoded at once, in a single arithmetic coding operation. A context for the
conditional encoding distribution is defined for each two-by-two group of
voxels, based on the information available about the occupancy of neighbor
voxels in the current and lower resolution layers of the octree. The CNN
estimates the probability distributions of occupancy patterns of all voxel
groups from one section in four phases. In each new phase the contexts are
updated with the occupancies encoded in the previous phase, and each phase
estimates the probabilities in parallel, providing a reasonable trade-off
between the parallelism of processing and the informativeness of the contexts.
The CNN training time is comparable to the time spent in the remaining encoding
steps, leading to competitive overall encoding times. Bitrates and
encoding-decoding times compare favorably with those of recently published
compression schemes.
Related papers
- High-Efficiency Neural Video Compression via Hierarchical Predictive Learning [27.41398149573729]
Enhanced Deep Hierarchical Video Compression-DHVC 2.0- introduces superior compression performance and impressive complexity efficiency.
Uses hierarchical predictive coding to transform each video frame into multiscale representations.
Supports transmission-friendly progressive decoding, making it particularly advantageous for networked video applications in the presence of packet loss.
arXiv Detail & Related papers (2024-10-03T15:40:58Z) - PINs: Progressive Implicit Networks for Multi-Scale Neural
Representations [68.73195473089324]
We propose a progressive positional encoding, exposing a hierarchical structure to incremental sets of frequency encodings.
Our model accurately reconstructs scenes with wide frequency bands and learns a scene representation at progressive level of detail.
Experiments on several 2D and 3D datasets show improvements in reconstruction accuracy, representational capacity and training speed compared to baselines.
arXiv Detail & Related papers (2022-02-09T20:33:37Z) - End-to-end Neural Video Coding Using a Compound Spatiotemporal
Representation [33.54844063875569]
We propose a hybrid motion compensation (HMC) method that adaptively combines the predictions generated by two approaches.
Specifically, we generate a compoundtemporal representation (STR) through a recurrent information aggregation (RIA) module.
We further design a one-to-many decoder pipeline to generate multiple predictions from the CSTR, including vector-based resampling, adaptive kernel-based resampling, compensation mode selection maps and texture enhancements.
arXiv Detail & Related papers (2021-08-05T19:43:32Z) - Dynamic Neural Representational Decoders for High-Resolution Semantic
Segmentation [98.05643473345474]
We propose a novel decoder, termed dynamic neural representational decoder (NRD)
As each location on the encoder's output corresponds to a local patch of the semantic labels, in this work, we represent these local patches of labels with compact neural networks.
This neural representation enables our decoder to leverage the smoothness prior in the semantic label space, and thus makes our decoder more efficient.
arXiv Detail & Related papers (2021-07-30T04:50:56Z) - Improved CNN-based Learning of Interpolation Filters for Low-Complexity
Inter Prediction in Video Coding [5.46121027847413]
This paper introduces a novel explainable neural network-based inter-prediction scheme.
A novel training framework enables each network branch to resemble a specific fractional shift.
When implemented in the context of the Versatile Video Coding (VVC) test model, 0.77%, 1.27% and 2.25% BD-rate savings can be achieved.
arXiv Detail & Related papers (2021-06-16T16:48:01Z) - Neural Network Modeling of Probabilities for Coding the Octree
Representation of Point Clouds [0.0]
This paper describes a novel point cloud compression algorithm that uses a neural network for estimating the coding probabilities for the occupancy status of voxels.
The proposed algorithms yield state-of-the-art results on benchmark datasets.
arXiv Detail & Related papers (2021-06-11T16:07:46Z) - Context- and Sequence-Aware Convolutional Recurrent Encoder for Neural
Machine Translation [2.729898906885749]
Existing models use recurrent neural networks to construct the encoder and decoder modules.
In alternative research, the recurrent networks were substituted by convolutional neural networks for capturing the syntactic structure in the input sentence.
We incorporate the goodness of both approaches by proposing a convolutional-recurrent encoder for capturing the context information.
arXiv Detail & Related papers (2021-01-11T17:03:52Z) - Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective
with Transformers [149.78470371525754]
We treat semantic segmentation as a sequence-to-sequence prediction task. Specifically, we deploy a pure transformer to encode an image as a sequence of patches.
With the global context modeled in every layer of the transformer, this encoder can be combined with a simple decoder to provide a powerful segmentation model, termed SEgmentation TRansformer (SETR)
SETR achieves new state of the art on ADE20K (50.28% mIoU), Pascal Context (55.83% mIoU) and competitive results on Cityscapes.
arXiv Detail & Related papers (2020-12-31T18:55:57Z) - Permute, Quantize, and Fine-tune: Efficient Compression of Neural
Networks [70.0243910593064]
Key to success of vector quantization is deciding which parameter groups should be compressed together.
In this paper we make the observation that the weights of two adjacent layers can be permuted while expressing the same function.
We then establish a connection to rate-distortion theory and search for permutations that result in networks that are easier to compress.
arXiv Detail & Related papers (2020-10-29T15:47:26Z) - Implicit Convex Regularizers of CNN Architectures: Convex Optimization
of Two- and Three-Layer Networks in Polynomial Time [70.15611146583068]
We study training of Convolutional Neural Networks (CNNs) with ReLU activations.
We introduce exact convex optimization with a complexity with respect to the number of data samples, the number of neurons, and data dimension.
arXiv Detail & Related papers (2020-06-26T04:47:20Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.