PVINet: Point-Voxel Interlaced Network for Point Cloud Compression
- URL: http://arxiv.org/abs/2509.01097v1
- Date: Mon, 01 Sep 2025 03:37:32 GMT
- Title: PVINet: Point-Voxel Interlaced Network for Point Cloud Compression
- Authors: Xuan Deng, Xingtao Wang, Xiandong Meng, Xiaopeng Fan, Debin Zhao,
- Abstract summary: In point cloud compression, the quality of a reconstructed point cloud relies on both the global structure and the local context.<n>We propose a point-voxel interlaced network (PVINet), which captures global structural features and local contextual features in parallel.<n>PVINet delivers competitive performance compared to state-of-the-art methods.
- Score: 83.74785652597248
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In point cloud compression, the quality of a reconstructed point cloud relies on both the global structure and the local context, with existing methods usually processing global and local information sequentially and lacking communication between these two types of information. In this paper, we propose a point-voxel interlaced network (PVINet), which captures global structural features and local contextual features in parallel and performs interactions at each scale to enhance feature perception efficiency. Specifically, PVINet contains a voxel-based encoder (Ev) for extracting global structural features and a point-based encoder (Ep) that models local contexts centered at each voxel. Particularly, a novel conditional sparse convolution is introduced, which applies point embeddings to dynamically customize kernels for voxel feature extraction, facilitating feature interactions from Ep to Ev. During decoding, a voxel-based decoder employs conditional sparse convolutions to incorporate point embeddings as guidance to reconstruct the point cloud. Experiments on benchmark datasets show that PVINet delivers competitive performance compared to state-of-the-art methods.
Related papers
- PVNet: Point-Voxel Interaction LiDAR Scene Upsampling Via Diffusion Models [57.02789948234898]
We propose PVNet, a diffusion model-based point-voxel interaction framework to perform LiDAR point cloud upsampling without dense supervision.<n>Specifically, we employ a sparse point cloud as the guiding condition and the synthesized point clouds derived from its nearby frames as the input.<n>In addition, we propose a point-voxel interaction module to integrate features from both points and voxels, which efficiently improves the environmental perception capability of each upsampled point.
arXiv Detail & Related papers (2025-08-23T14:55:03Z) - PointCFormer: a Relation-based Progressive Feature Extraction Network for Point Cloud Completion [19.503392612245474]
Point cloud completion aims to reconstruct the complete 3D shape from incomplete point clouds.<n>We introduce PointCFormer, a transformer framework optimized for robust global retention and precise local detail capture.<n>PointCFormer demonstrates state-of-the-art performance on several widely used benchmarks.
arXiv Detail & Related papers (2024-12-11T14:37:21Z) - DiffCom: Decoupled Sparse Priors Guided Diffusion Compression for Point Clouds [54.96190721255167]
Lossy compression relies on an autoencoder to transform a point cloud into latent points for storage.<n>We propose a diffusion-based framework guided by sparse priors that achieves high reconstruction quality, especially at lows.
arXiv Detail & Related papers (2024-11-21T05:41:35Z) - Point Cloud Compression with Implicit Neural Representations: A Unified Framework [54.119415852585306]
We present a pioneering point cloud compression framework capable of handling both geometry and attribute components.
Our framework utilizes two coordinate-based neural networks to implicitly represent a voxelized point cloud.
Our method exhibits high universality when contrasted with existing learning-based techniques.
arXiv Detail & Related papers (2024-05-19T09:19:40Z) - Collect-and-Distribute Transformer for 3D Point Cloud Analysis [82.03517861433849]
We propose a new transformer network equipped with a collect-and-distribute mechanism to communicate short- and long-range contexts of point clouds.
Results show the effectiveness of the proposed CDFormer, delivering several new state-of-the-art performances on point cloud classification and segmentation tasks.
arXiv Detail & Related papers (2023-06-02T03:48:45Z) - VTPNet for 3D deep learning on point cloud [10.470127366415813]
Voxel-Transformer-Point (VTP) Block for extracting local and global features of point clouds.
VTP combines the advantages of voxel-based, point-based and Transformer-based methods.
Experiments indicate that VTPNet has good performance in 3D point cloud learning.
arXiv Detail & Related papers (2023-05-10T13:07:46Z) - SemAffiNet: Semantic-Affine Transformation for Point Cloud Segmentation [94.11915008006483]
We propose SemAffiNet for point cloud semantic segmentation.
We conduct extensive experiments on the ScanNetV2 and NYUv2 datasets.
arXiv Detail & Related papers (2022-05-26T17:00:23Z) - Point cloud completion on structured feature map with feedback network [28.710494879042002]
We propose FSNet, a feature structuring module that can adaptively aggregate point-wise features into a 2D structured feature map.
A 2D convolutional neural network is adopted to decode feature maps from FSNet into a coarse and complete point cloud.
A point cloud upsampling network is used to generate dense point cloud from the partial input and the coarse intermediate output.
arXiv Detail & Related papers (2022-02-17T10:59:40Z) - Fast Point Voxel Convolution Neural Network with Selective Feature
Fusion for Point Cloud Semantic Segmentation [7.557684072809662]
We present a novel lightweight convolutional neural network for point cloud analysis.
Our method operates on the entire point sets without sampling and achieves good performances efficiently.
arXiv Detail & Related papers (2021-09-23T19:39:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.