Point Transformer for Shape Classification and Retrieval of 3D and ALS
Roof PointClouds
- URL: http://arxiv.org/abs/2011.03921v2
- Date: Sat, 20 Feb 2021 08:54:37 GMT
- Title: Point Transformer for Shape Classification and Retrieval of 3D and ALS
Roof PointClouds
- Authors: Dimple A Shajahan, Mukund Varma T and Ramanathan Muthuganapathy
- Abstract summary: This paper proposes a fully attentional model - em Point Transformer, for deriving a rich point cloud representation.
The model's shape classification and retrieval performance are evaluated on a large-scale urban dataset - RoofN3D and a standard benchmark dataset ModelNet40.
The proposed method outperforms other state-of-the-art models in the RoofN3D dataset, gives competitive results in the ModelNet40 benchmark, and showcases high robustness to various unseen point corruptions.
- Score: 3.3744638598036123
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The success of deep learning methods led to significant breakthroughs in 3-D
point cloud processing tasks with applications in remote sensing. Existing
methods utilize convolutions that have some limitations, as they assume a
uniform input distribution and cannot learn long-range dependencies. Recent
works have shown that adding attention in conjunction with these methods
improves performance. This raises a question: can attention layers completely
replace convolutions? This paper proposes a fully attentional model - {\em
Point Transformer}, for deriving a rich point cloud representation. The model's
shape classification and retrieval performance are evaluated on a large-scale
urban dataset - RoofN3D and a standard benchmark dataset ModelNet40. Extensive
experiments are conducted to test the model's robustness to unseen point
corruptions for analyzing its effectiveness on real datasets. The proposed
method outperforms other state-of-the-art models in the RoofN3D dataset, gives
competitive results in the ModelNet40 benchmark, and showcases high robustness
to various unseen point corruptions. Furthermore, the model is highly memory
and space efficient when compared to other methods.
Related papers
- Foundation Model-Powered 3D Few-Shot Class Incremental Learning via Training-free Adaptor [9.54964908165465]
This paper introduces a new method to tackle the Few-Shot Continual Incremental Learning problem in 3D point cloud environments.
We leverage a foundational 3D model trained extensively on point cloud data.
Our approach uses a dual cache system: first, it uses previous test samples based on how confident the model was in its predictions to prevent forgetting, and second, it includes a small number of new task samples to prevent overfitting.
arXiv Detail & Related papers (2024-10-11T20:23:00Z) - Transfer Learning with Point Transformers [3.678615604632945]
Point Transformers are state-of-the-art models for classification, segmentation, and detection on Point Cloud data.
We explore two things: classification performance of these attention based networks on ModelNet10 dataset and then, we use the trained model to classify 3D MNIST dataset after finetuning.
arXiv Detail & Related papers (2024-04-01T01:23:58Z) - ModelNet-O: A Large-Scale Synthetic Dataset for Occlusion-Aware Point
Cloud Classification [28.05358017259757]
We propose ModelNet-O, a large-scale synthetic dataset of 123,041 samples.
ModelNet-O emulates real-world point clouds with self-occlusion caused by scanning from monocular cameras.
We propose a robust point cloud processing method called PointMLS.
arXiv Detail & Related papers (2024-01-16T08:54:21Z) - FILP-3D: Enhancing 3D Few-shot Class-incremental Learning with
Pre-trained Vision-Language Models [62.663113296987085]
Few-shot class-incremental learning aims to mitigate the catastrophic forgetting issue when a model is incrementally trained on limited data.
We introduce two novel components: the Redundant Feature Eliminator (RFE) and the Spatial Noise Compensator (SNC)
Considering the imbalance in existing 3D datasets, we also propose new evaluation metrics that offer a more nuanced assessment of a 3D FSCIL model.
arXiv Detail & Related papers (2023-12-28T14:52:07Z) - StarNet: Style-Aware 3D Point Cloud Generation [82.30389817015877]
StarNet is able to reconstruct and generate high-fidelity and even 3D point clouds using a mapping network.
Our framework achieves comparable state-of-the-art performance on various metrics in the point cloud reconstruction and generation tasks.
arXiv Detail & Related papers (2023-03-28T08:21:44Z) - 3DVerifier: Efficient Robustness Verification for 3D Point Cloud Models [17.487852393066458]
Existing verification method for point cloud model is time-expensive and computationally unattainable on large networks.
We propose 3DVerifier to tackle both challenges by adopting a linear relaxation function to bound the multiplication layer and combining forward and backward propagation.
Our approach achieves an orders-of-magnitude improvement in verification efficiency for the large network, and the obtained certified bounds are also significantly tighter than the state-of-the-art verifiers.
arXiv Detail & Related papers (2022-07-15T15:31:16Z) - Point Discriminative Learning for Unsupervised Representation Learning
on 3D Point Clouds [54.31515001741987]
We propose a point discriminative learning method for unsupervised representation learning on 3D point clouds.
We achieve this by imposing a novel point discrimination loss on the middle level and global level point features.
Our method learns powerful representations and achieves new state-of-the-art performance.
arXiv Detail & Related papers (2021-08-04T15:11:48Z) - Learning Semantic Segmentation of Large-Scale Point Clouds with Random
Sampling [52.464516118826765]
We introduce RandLA-Net, an efficient and lightweight neural architecture to infer per-point semantics for large-scale point clouds.
The key to our approach is to use random point sampling instead of more complex point selection approaches.
Our RandLA-Net can process 1 million points in a single pass up to 200x faster than existing approaches.
arXiv Detail & Related papers (2021-07-06T05:08:34Z) - Secrets of 3D Implicit Object Shape Reconstruction in the Wild [92.5554695397653]
Reconstructing high-fidelity 3D objects from sparse, partial observation is crucial for various applications in computer vision, robotics, and graphics.
Recent neural implicit modeling methods show promising results on synthetic or dense datasets.
But, they perform poorly on real-world data that is sparse and noisy.
This paper analyzes the root cause of such deficient performance of a popular neural implicit model.
arXiv Detail & Related papers (2021-01-18T03:24:48Z) - 2nd Place Scheme on Action Recognition Track of ECCV 2020 VIPriors
Challenges: An Efficient Optical Flow Stream Guided Framework [57.847010327319964]
We propose a data-efficient framework that can train the model from scratch on small datasets.
Specifically, by introducing a 3D central difference convolution operation, we proposed a novel C3D neural network-based two-stream framework.
It is proved that our method can achieve a promising result even without a pre-trained model on large scale datasets.
arXiv Detail & Related papers (2020-08-10T09:50:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.