Neural Network Modeling of Probabilities for Coding the Octree
Representation of Point Clouds
- URL: http://arxiv.org/abs/2106.06482v2
- Date: Mon, 14 Jun 2021 13:28:20 GMT
- Title: Neural Network Modeling of Probabilities for Coding the Octree
Representation of Point Clouds
- Authors: Emre Can Kaya, Ioan Tabus
- Abstract summary: This paper describes a novel point cloud compression algorithm that uses a neural network for estimating the coding probabilities for the occupancy status of voxels.
The proposed algorithms yield state-of-the-art results on benchmark datasets.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper describes a novel lossless point cloud compression algorithm that
uses a neural network for estimating the coding probabilities for the occupancy
status of voxels, depending on wide three dimensional contexts around the voxel
to be encoded. The point cloud is represented as an octree, with each
resolution layer being sequentially encoded and decoded using arithmetic
coding, starting from the lowest resolution, until the final resolution is
reached. The occupancy probability of each voxel of the splitting pattern at
each node of the octree is modeled by a neural network, having at its input the
already encoded occupancy status of several octree nodes (belonging to the past
and current resolutions), corresponding to a 3D context surrounding the node to
be encoded. The algorithm has a fast and a slow version, the fast version
selecting differently several voxels of the context, which allows an increased
parallelization by sending larger batches of templates to be estimated by the
neural network, at both encoder and decoder. The proposed algorithms yield
state-of-the-art results on benchmark datasets. The implementation will be made
available at https://github.com/marmus12/nnctx
Related papers
- Multi-Scale Node Embeddings for Graph Modeling and Generation [0.0]
node embedding algorithms take a graph as input and encode its structure onto output vectors that represent nodes in a geometric space.
Two apparently unrelated limitations affect these algorithms.
We define a multiscale node embedding method that ensures statistical consistency of the embedding vector of a block-node with the sum of the embedding vectors of its constituent nodes.
arXiv Detail & Related papers (2024-12-05T17:12:45Z) - Learning Neural Volumetric Field for Point Cloud Geometry Compression [13.691147541041804]
We propose to code the geometry of a given point cloud by learning a neural field.
We divide the entire space into small cubes and represent each non-empty cube by a neural network and an input latent code.
The network is shared among all the cubes in a single frame or multiple frames, to exploit the spatial and temporal redundancy.
arXiv Detail & Related papers (2022-12-11T19:55:24Z) - Variable Bitrate Neural Fields [75.24672452527795]
We present a dictionary method for compressing feature grids, reducing their memory consumption by up to 100x.
We formulate the dictionary optimization as a vector-quantized auto-decoder problem which lets us learn end-to-end discrete neural representations in a space where no direct supervision is available.
arXiv Detail & Related papers (2022-06-15T17:58:34Z) - Lossless Compression of Point Cloud Sequences Using Sequence Optimized
CNN Models [0.0]
We propose a new paradigm for encoding the geometry of point cloud sequences, where the convolutional neural network estimates the encoding distributions.
We adopt lightweight CNN structures, we perform training as part of the encoding process, and the CNN parameters are transmitted as part of the bitstream.
arXiv Detail & Related papers (2022-06-02T20:46:05Z) - SoftPool++: An Encoder-Decoder Network for Point Cloud Completion [93.54286830844134]
We propose a novel convolutional operator for the task of point cloud completion.
The proposed operator does not require any max-pooling or voxelization operation.
We show that our approach achieves state-of-the-art performance in shape completion at low and high resolutions.
arXiv Detail & Related papers (2022-05-08T15:31:36Z) - Instant Neural Graphics Primitives with a Multiresolution Hash Encoding [67.33850633281803]
We present a versatile new input encoding that permits the use of a smaller network without sacrificing quality.
A small neural network is augmented by a multiresolution hash table of trainable feature vectors whose values are optimized through a gradient descent.
We achieve a combined speed of several orders of magnitude, enabling training of high-quality neural graphics primitives in a matter of seconds.
arXiv Detail & Related papers (2022-01-16T07:22:47Z) - Dynamic Neural Representational Decoders for High-Resolution Semantic
Segmentation [98.05643473345474]
We propose a novel decoder, termed dynamic neural representational decoder (NRD)
As each location on the encoder's output corresponds to a local patch of the semantic labels, in this work, we represent these local patches of labels with compact neural networks.
This neural representation enables our decoder to leverage the smoothness prior in the semantic label space, and thus makes our decoder more efficient.
arXiv Detail & Related papers (2021-07-30T04:50:56Z) - Lossless Coding of Point Cloud Geometry using a Deep Generative Model [11.69103847045569]
Method adaptively partitions a point cloud into multiple voxel block sizes.
Deep auto-regressive generative model estimates occupancy probability of each voxel.
We employ estimated probabilities to code efficiently a block using a context-based arithmetic coder.
arXiv Detail & Related papers (2021-07-01T12:20:22Z) - Learning-based lossless compression of 3D point cloud geometry [11.69103847045569]
encoder operates in a hybrid mode, mixing octree and voxel-based coding.
Our method outperforms the state-of-the-art MPEG G-PCC standard with average rate savings of 28%.
arXiv Detail & Related papers (2020-11-30T11:27:16Z) - Permute, Quantize, and Fine-tune: Efficient Compression of Neural
Networks [70.0243910593064]
Key to success of vector quantization is deciding which parameter groups should be compressed together.
In this paper we make the observation that the weights of two adjacent layers can be permuted while expressing the same function.
We then establish a connection to rate-distortion theory and search for permutations that result in networks that are easier to compress.
arXiv Detail & Related papers (2020-10-29T15:47:26Z) - Refinement of Predicted Missing Parts Enhance Point Cloud Completion [62.997667081978825]
Point cloud completion is the task of predicting complete geometry from partial observations using a point set representation for a 3D shape.
Previous approaches propose neural networks to directly estimate the whole point cloud through encoder-decoder models fed by the incomplete point set.
This paper proposes an end-to-end neural network architecture that focuses on computing the missing geometry and merging the known input and the predicted point cloud.
arXiv Detail & Related papers (2020-10-08T22:01:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.