POEM: 1-bit Point-wise Operations based on Expectation-Maximization for
Efficient Point Cloud Processing
- URL: http://arxiv.org/abs/2111.13386v1
- Date: Fri, 26 Nov 2021 09:45:01 GMT
- Title: POEM: 1-bit Point-wise Operations based on Expectation-Maximization for
Efficient Point Cloud Processing
- Authors: Sheng Xu, Yanjing Li, Junhe Zhao, Baochang Zhang, Guodong Guo
- Abstract summary: We introduce point-wise operations based on Expectation-Maximization into BNNs for efficient point cloud processing.
Our POEM surpasses existing the state-of-the-art binary point cloud networks by a significant margin, up to 6.7 %.
- Score: 53.74076015905961
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Real-time point cloud processing is fundamental for lots of computer vision
tasks, while still challenged by the computational problem on resource-limited
edge devices. To address this issue, we implement XNOR-Net-based binary neural
networks (BNNs) for an efficient point cloud processing, but its performance is
severely suffered due to two main drawbacks, Gaussian-distributed weights and
non-learnable scale factor. In this paper, we introduce point-wise operations
based on Expectation-Maximization (POEM) into BNNs for efficient point cloud
processing. The EM algorithm can efficiently constrain weights for a robust
bi-modal distribution. We lead a well-designed reconstruction loss to calculate
learnable scale factors to enhance the representation capacity of 1-bit
fully-connected (Bi-FC) layers. Extensive experiments demonstrate that our POEM
surpasses existing the state-of-the-art binary point cloud networks by a
significant margin, up to 6.7 %.
Related papers
- ReActXGB: A Hybrid Binary Convolutional Neural Network Architecture for Improved Performance and Computational Efficiency [0.0]
We propose a hybrid model named ReActXGB, where we replace the fully convolutional layer of ReActNet-A with XGBoost.
This modification targets to narrow the performance gap between BCNNs and real-valued networks while maintaining lower computational costs.
arXiv Detail & Related papers (2024-05-11T16:38:50Z) - FBPT: A Fully Binary Point Transformer [12.373066597900127]
This paper presents a novel Fully Binary Point Cloud Transformer (FBPT) model which has the potential to be widely applied and expanded in the fields of robotics and mobile devices.
By compressing the weights and activations of a 32-bit full-precision network to 1-bit binary values, the proposed binary point cloud Transformer network significantly reduces the storage footprint and computational resource requirements.
The primary focus of this paper is on addressing the performance degradation issue caused by the use of binary point cloud Transformer modules.
arXiv Detail & Related papers (2024-03-15T03:45:10Z) - Projected Stochastic Gradient Descent with Quantum Annealed Binary Gradients [51.82488018573326]
We present QP-SBGD, a novel layer-wise optimiser tailored towards training neural networks with binary weights.
BNNs reduce the computational requirements and energy consumption of deep learning models with minimal loss in accuracy.
Our algorithm is implemented layer-wise, making it suitable to train larger networks on resource-limited quantum hardware.
arXiv Detail & Related papers (2023-10-23T17:32:38Z) - Efficient Deep Spiking Multi-Layer Perceptrons with Multiplication-Free Inference [13.924924047051782]
Deep convolution architectures for Spiking Neural Networks (SNNs) have significantly enhanced image classification performance and reduced computational burdens.
This research explores a new pathway, drawing inspiration from the progress made in Multi-Layer Perceptrons (MLPs)
We propose an innovative spiking architecture that uses batch normalization to retain MFI compatibility.
We establish an efficient multi-stage spiking network that blends effectively global receptive fields with local feature extraction.
arXiv Detail & Related papers (2023-06-21T16:52:20Z) - Binarizing Sparse Convolutional Networks for Efficient Point Cloud
Analysis [93.55896765176414]
We propose binary sparse convolutional networks called BSC-Net for efficient point cloud analysis.
We employ the differentiable search strategies to discover the optimal opsitions for active site matching in the shifted sparse convolution.
Our BSC-Net achieves significant improvement upon our srtong baseline and outperforms the state-of-the-art network binarization methods.
arXiv Detail & Related papers (2023-03-27T13:47:06Z) - Recurrent Bilinear Optimization for Binary Neural Networks [58.972212365275595]
BNNs neglect the intrinsic bilinear relationship of real-valued weights and scale factors.
Our work is the first attempt to optimize BNNs from the bilinear perspective.
We obtain robust RBONNs, which show impressive performance over state-of-the-art BNNs on various models and datasets.
arXiv Detail & Related papers (2022-09-04T06:45:33Z) - FTBNN: Rethinking Non-linearity for 1-bit CNNs and Going Beyond [23.5996182207431]
We show that binarized convolution process owns an increasing linearity towards the target of minimizing such error, which in turn hampers BNN's discriminative ability.
We re-investigate and tune proper non-linear modules to fix that contradiction, leading to a strong baseline which achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-10-19T08:11:48Z) - BiPointNet: Binary Neural Network for Point Clouds [73.07852523426224]
BiPointNet is the first model binarization approach for efficient deep learning on point clouds.
We show that BiPointNet gives an impressive 14.7x speedup and 18.9x storage saving on real-world resource-constrained devices.
arXiv Detail & Related papers (2020-10-12T07:54:51Z) - Widening and Squeezing: Towards Accurate and Efficient QNNs [125.172220129257]
Quantization neural networks (QNNs) are very attractive to the industry because their extremely cheap calculation and storage overhead, but their performance is still worse than that of networks with full-precision parameters.
Most of existing methods aim to enhance performance of QNNs especially binary neural networks by exploiting more effective training techniques.
We address this problem by projecting features in original full-precision networks to high-dimensional quantization features.
arXiv Detail & Related papers (2020-02-03T04:11:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.