Irregularly Tabulated MLP for Fast Point Feature Embedding
- URL: http://arxiv.org/abs/2011.09852v1
- Date: Fri, 13 Nov 2020 04:15:57 GMT
- Title: Irregularly Tabulated MLP for Fast Point Feature Embedding
- Authors: Yusuke Sekikawa, Teppei Suzuki
- Abstract summary: We propose a new framework that uses a pair of multi-layer perceptrons (MLP) and a lookup table (LUT) to transform point-coordinate inputs into high-dimensional features.
LUTIMLP also provides significant speedup for Jacobian of the embedding function.
- Score: 13.218995242910497
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Aiming at drastic speedup for point-feature embeddings at test time, we
propose a new framework that uses a pair of multi-layer perceptrons (MLP) and a
lookup table (LUT) to transform point-coordinate inputs into high-dimensional
features. When compared with PointNet's feature embedding part realized by MLP
that requires millions of dot products, the proposed framework at test time
requires no such layers of matrix-vector products but requires only looking up
the nearest entities from the tabulated MLP followed by interpolation, defined
over discrete inputs on a 3D lattice that is substantially arranged
irregularly. We call this framework LUTI-MLP: LUT Interpolation ML that
provides a way to train end-to-end irregularly tabulated MLP coupled to a LUT
in a specific manner without the need for any approximation at test time.
LUTI-MLP also provides significant speedup for Jacobian computation of the
embedding function wrt global pose coordinate on Lie algebra $\mathfrak{se}(3)$
at test time, which could be used for point-set registration problems. After
extensive evaluation using the ModelNet40, we confirmed that the LUTI-MLP even
with a small (e.g., $4^3$) lattice yields performance comparable to that of the
MLP while achieving significant speedup: $100\times$ for the embedding,
$12\times$ for the approximate Jacobian, and $860\times$ for the canonical
Jacobian.
Related papers
- SiT-MLP: A Simple MLP with Point-wise Topology Feature Learning for Skeleton-based Action Recognition [9.673505408890435]
Graph networks (GCNs) have achieved remarkable performance in skeleton-based action recognition.
Previous GCN-based methods rely on elaborate human priors excessively and construct complex feature aggregation mechanisms.
We propose a novel model, SiT-MLP, for skeleton-based action recognition in this work.
arXiv Detail & Related papers (2023-08-30T13:20:54Z) - Tuning Pre-trained Model via Moment Probing [62.445281364055795]
We propose a novel Moment Probing (MP) method to explore the potential of LP.
MP performs a linear classification head based on the mean of final features.
Our MP significantly outperforms LP and is competitive with counterparts at less training cost.
arXiv Detail & Related papers (2023-07-21T04:15:02Z) - NTK-approximating MLP Fusion for Efficient Language Model Fine-tuning [40.994306592119266]
Fine-tuning a pre-trained language model (PLM) emerges as the predominant strategy in many natural language processing applications.
Some general approaches (e.g. quantization and distillation) have been widely studied to reduce the compute/memory of PLM fine-tuning.
We propose to coin a lightweight PLM through NTK-approximating modules in fusion.
arXiv Detail & Related papers (2023-07-18T03:12:51Z) - TriMLP: Revenge of a MLP-like Architecture in Sequential Recommendation [23.32537260687907]
We present a sequential-like architecture for sequential recommendation, namely TriMLP, with a novel Triangular Mixer for cross-token communications.
In designing Triangular Mixer, we simplify the cross-token operation inascii as the basic matrix multiplication, and drop the lower-triangle neurons of the weight matrix to block the anti-chronological order connections from future tokens.
arXiv Detail & Related papers (2023-05-24T03:32:31Z) - Parameterization of Cross-Token Relations with Relative Positional
Encoding for Vision MLP [52.25478388220691]
Vision multi-layer perceptrons (MLPs) have shown promising performance in computer vision tasks.
They use token-mixing layers to capture cross-token interactions, as opposed to the multi-head self-attention mechanism used by Transformers.
We propose a new positional spacial gating unit (PoSGU) to efficiently encode the cross-token relations for token mixing.
arXiv Detail & Related papers (2022-07-15T04:18:06Z) - ReLU Fields: The Little Non-linearity That Could [62.228229880658404]
We investigate what is the smallest change to grid-based representations that allows for retaining the high fidelity result ofs.
We show that such an approach becomes competitive with the state-of-the-art.
arXiv Detail & Related papers (2022-05-22T13:42:31Z) - Rethinking Network Design and Local Geometry in Point Cloud: A Simple
Residual MLP Framework [55.40001810884942]
We introduce a pure residual network, called PointMLP, which integrates no sophisticated local geometrical extractors but still performs very competitively.
On the real-world ScanObjectNN dataset, our method even surpasses the prior best method by 3.3% accuracy.
Compared to most recent CurveNet, PointMLP trains 2x faster, tests 7x faster, and is more accurate on ModelNet40 benchmark.
arXiv Detail & Related papers (2022-02-15T01:39:07Z) - Hire-MLP: Vision MLP via Hierarchical Rearrangement [58.33383667626998]
Hire-MLP is a simple yet competitive vision architecture via rearrangement.
The proposed Hire-MLP architecture is built with simple channel-mixing operations, thus enjoys high flexibility and inference speed.
Experiments show that our Hire-MLP achieves state-of-the-art performance on the ImageNet-1K benchmark.
arXiv Detail & Related papers (2021-08-30T16:11:04Z) - CycleMLP: A MLP-like Architecture for Dense Prediction [26.74203747156439]
CycleMLP is a versatile backbone for visual recognition and dense predictions.
It can cope with various image sizes and achieves linear computational complexity to image size by using local windows.
CycleMLP aims to provide a competitive baseline on object detection, instance segmentation, and semantic segmentation for models.
arXiv Detail & Related papers (2021-07-21T17:23:06Z) - AS-MLP: An Axial Shifted MLP Architecture for Vision [50.11765148947432]
An Axial Shifted architecture (AS-MLP) is proposed in this paper.
By axially shifting channels of the feature map, AS-MLP is able to obtain the information flow from different directions.
With the proposed AS-MLP architecture, our model obtains 83.3% Top-1 accuracy with 88M parameters and 15.2 GFLOPs on the ImageNet-1K dataset.
arXiv Detail & Related papers (2021-07-18T08:56:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.