Volumetric Attribute Compression for 3D Point Clouds using Feedforward
Network with Geometric Attention
- URL: http://arxiv.org/abs/2304.00335v1
- Date: Sat, 1 Apr 2023 15:24:12 GMT
- Title: Volumetric Attribute Compression for 3D Point Clouds using Feedforward
Network with Geometric Attention
- Authors: Tam Thuc Do, Philip A. Chou, Gene Cheung
- Abstract summary: We propose a feedforward linear network that implements higher-order B-spline bases spanning function spaces without eigendecomposition.
We show that the number of layers in the normalization at the encoder is equivalent to the number of terms in an inverse Taylor series.
- Score: 36.41214415449853
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We study 3D point cloud attribute compression using a volumetric approach:
given a target volumetric attribute function $f : \mathbb{R}^3 \rightarrow
\mathbb{R}$, we quantize and encode parameter vector $\theta$ that
characterizes $f$ at the encoder, for reconstruction
$f_{\hat{\theta}}(\mathbf{x})$ at known 3D points $\mathbf{x}$'s at the
decoder. Extending a previous work Region Adaptive Hierarchical Transform
(RAHT) that employs piecewise constant functions to span a nested sequence of
function spaces, we propose a feedforward linear network that implements
higher-order B-spline bases spanning function spaces without
eigen-decomposition. Feedforward network architecture means that the system is
amenable to end-to-end neural learning. The key to our network is space-varying
convolution, similar to a graph operator, whose weights are computed from the
known 3D geometry for normalization. We show that the number of layers in the
normalization at the encoder is equivalent to the number of terms in a matrix
inverse Taylor series. Experimental results on real-world 3D point clouds show
up to 2-3 dB gain over RAHT in energy compaction and 20-30% bitrate reduction.
Related papers
- Deep Unrolling of Sparsity-Induced RDO for 3D Point Cloud Attribute Coding [24.375903431917163]
We study the problem of lossy attribute compression in a multi-resolution B-spline projection framework.<n>A target continuous 3D attribute function is first projected onto a sequence of nested subspaces.<n>For a chosen coarse-to-fine predictor, the coefficients are then adjusted to account for the prediction from a lower-resolution to a higher-resolution.
arXiv Detail & Related papers (2025-09-10T15:23:21Z) - ReMiDi: Reconstruction of Microstructure Using a Differentiable Diffusion MRI Simulator [0.602276990341246]
ReMiDi is a novel method for inferring neuronal microstructure as arbitrary 3D meshes using a differentiable diffusion Magnetic Resonance Imaging (dMRI) simulator.
We present an end-to-end differentiable pipeline that simulates signals that can be tuned to match a reference signal.
We demonstrate the ability to reconstruct microstructures of arbitrary shapes represented by finite-element meshes, with a focus on axonal geometries found in the brain white matter.
arXiv Detail & Related papers (2025-02-04T04:03:08Z) - Implicit Hypersurface Approximation Capacity in Deep ReLU Networks [0.0]
We develop a geometric approximation theory for deep feed-forward neural networks with ReLU activations.
We show that a deep fully-connected ReLU network of width $d+1$ can implicitly construct an approximation as its zero contour.
arXiv Detail & Related papers (2024-07-04T11:34:42Z) - Learning Hierarchical Polynomials with Three-Layer Neural Networks [56.71223169861528]
We study the problem of learning hierarchical functions over the standard Gaussian distribution with three-layer neural networks.
For a large subclass of degree $k$s $p$, a three-layer neural network trained via layerwise gradientp descent on the square loss learns the target $h$ up to vanishing test error.
This work demonstrates the ability of three-layer neural networks to learn complex features and as a result, learn a broad class of hierarchical functions.
arXiv Detail & Related papers (2023-11-23T02:19:32Z) - Learned Nonlinear Predictor for Critically Sampled 3D Point Cloud
Attribute Compression [24.001318485207207]
We study 3D point cloud compression via a decoder approach.
In this paper, we study predicting $f_l*$ at level $l+1$ given $f_l*$ $l$ and encoding of $G_l*$ for the $p=1$ case.
arXiv Detail & Related papers (2023-11-22T17:26:54Z) - On Expressivity of Height in Neural Networks [29.49793694185358]
We call a neural network characterized by width, depth, and height a 3D network.
We show via bound estimation and explicit construction that given the same number of neurons and parameters, a 3D ReLU network of width $W$, depth $K$, and height $H$ has greater expressive power than a 2D network of width $Htimes W$ and depth $K$.
arXiv Detail & Related papers (2023-05-11T11:54:36Z) - Learning Neural Volumetric Field for Point Cloud Geometry Compression [13.691147541041804]
We propose to code the geometry of a given point cloud by learning a neural field.
We divide the entire space into small cubes and represent each non-empty cube by a neural network and an input latent code.
The network is shared among all the cubes in a single frame or multiple frames, to exploit the spatial and temporal redundancy.
arXiv Detail & Related papers (2022-12-11T19:55:24Z) - LVAC: Learned Volumetric Attribute Compression for Point Clouds using
Coordinate Based Networks [21.6781972169876]
We consider the attributes of a point cloud as samples of a vector-valued volumetric function at discrete positions.
We model the volumetric function by tiling space into blocks, and representing the function over each block by shifts of a coordinate-based, or implicit, neural network.
We represent the latent vectors using coefficients of the region-adaptive hierarchical transform (RAHT) used in the geometry-based point cloud G-PCC.
arXiv Detail & Related papers (2021-11-17T09:11:09Z) - Learning Deformable Tetrahedral Meshes for 3D Reconstruction [78.0514377738632]
3D shape representations that accommodate learning-based 3D reconstruction are an open problem in machine learning and computer graphics.
Previous work on neural 3D reconstruction demonstrated benefits, but also limitations, of point cloud, voxel, surface mesh, and implicit function representations.
We introduce Deformable Tetrahedral Meshes (DefTet) as a particular parameterization that utilizes volumetric tetrahedral meshes for the reconstruction problem.
arXiv Detail & Related papers (2020-11-03T02:57:01Z) - Learning Local Neighboring Structure for Robust 3D Shape Representation [143.15904669246697]
Representation learning for 3D meshes is important in many computer vision and graphics applications.
We propose a local structure-aware anisotropic convolutional operation (LSA-Conv)
Our model produces significant improvement in 3D shape reconstruction compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-04-21T13:40:03Z) - Implicit Functions in Feature Space for 3D Shape Reconstruction and
Completion [53.885984328273686]
Implicit Feature Networks (IF-Nets) deliver continuous outputs, can handle multiple topologies, and complete shapes for missing or sparse input data.
IF-Nets clearly outperform prior work in 3D object reconstruction in ShapeNet, and obtain significantly more accurate 3D human reconstructions.
arXiv Detail & Related papers (2020-03-03T11:14:29Z) - PUGeo-Net: A Geometry-centric Network for 3D Point Cloud Upsampling [103.09504572409449]
We propose a novel deep neural network based method, called PUGeo-Net, to generate uniform dense point clouds.
Thanks to its geometry-centric nature, PUGeo-Net works well for both CAD models with sharp features and scanned models with rich geometric details.
arXiv Detail & Related papers (2020-02-24T14:13:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.