PointCAT: Cross-Attention Transformer for point cloud
- URL: http://arxiv.org/abs/2304.03012v1
- Date: Thu, 6 Apr 2023 11:58:18 GMT
- Title: PointCAT: Cross-Attention Transformer for point cloud
- Authors: Xincheng Yang, Mingze Jin, Weiji He, Qian Chen
- Abstract summary: We present Point Cross-Attention Transformer (PointCAT), a novel end-to-end network architecture.
Our approach combines multi-scale features via two seprate cross-attention transformer branches.
Our method outperforms or achieves comparable performance to several approaches in shape classification, part segmentation and semantic segmentation tasks.
- Score: 1.3176016397292067
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transformer-based models have significantly advanced natural language
processing and computer vision in recent years. However, due to the irregular
and disordered structure of point cloud data, transformer-based models for 3D
deep learning are still in their infancy compared to other methods. In this
paper we present Point Cross-Attention Transformer (PointCAT), a novel
end-to-end network architecture using cross-attentions mechanism for point
cloud representing. Our approach combines multi-scale features via two seprate
cross-attention transformer branches. To reduce the computational increase
brought by multi-branch structure, we further introduce an efficient model for
shape classification, which only process single class token of one branch as a
query to calculate attention map with the other. Extensive experiments
demonstrate that our method outperforms or achieves comparable performance to
several approaches in shape classification, part segmentation and semantic
segmentation tasks.
Related papers
- Boosting Cross-Domain Point Classification via Distilling Relational Priors from 2D Transformers [59.0181939916084]
Traditional 3D networks mainly focus on local geometric details and ignore the topological structure between local geometries.
We propose a novel Priors Distillation (RPD) method to extract priors from the well-trained transformers on massive images.
Experiments on the PointDA-10 and the Sim-to-Real datasets verify that the proposed method consistently achieves the state-of-the-art performance of UDA for point cloud classification.
arXiv Detail & Related papers (2024-07-26T06:29:09Z) - AdaPoinTr: Diverse Point Cloud Completion with Adaptive Geometry-Aware
Transformers [94.11915008006483]
We present a new method that reformulates point cloud completion as a set-to-set translation problem.
We design a new model, called PoinTr, which adopts a Transformer encoder-decoder architecture for point cloud completion.
Our method attains 6.53 CD on PCN, 0.81 CD on ShapeNet-55 and 0.392 MMD on real-world KITTI.
arXiv Detail & Related papers (2023-01-11T16:14:12Z) - CloudAttention: Efficient Multi-Scale Attention Scheme For 3D Point
Cloud Learning [81.85951026033787]
We set transformers in this work and incorporate them into a hierarchical framework for shape classification and part and scene segmentation.
We also compute efficient and dynamic global cross attentions by leveraging sampling and grouping at each iteration.
The proposed hierarchical model achieves state-of-the-art shape classification in mean accuracy and yields results on par with the previous segmentation methods.
arXiv Detail & Related papers (2022-07-31T21:39:15Z) - SemAffiNet: Semantic-Affine Transformation for Point Cloud Segmentation [94.11915008006483]
We propose SemAffiNet for point cloud semantic segmentation.
We conduct extensive experiments on the ScanNetV2 and NYUv2 datasets.
arXiv Detail & Related papers (2022-05-26T17:00:23Z) - Semantic Correspondence with Transformers [68.37049687360705]
We propose Cost Aggregation with Transformers (CATs) to find dense correspondences between semantically similar images.
We include appearance affinity modelling to disambiguate the initial correlation maps and multi-level aggregation.
We conduct experiments to demonstrate the effectiveness of the proposed model over the latest methods and provide extensive ablation studies.
arXiv Detail & Related papers (2021-06-04T14:39:03Z) - Point Cloud Learning with Transformer [2.3204178451683264]
We introduce a novel framework, called Multi-level Multi-scale Point Transformer (MLMSPT)
Specifically, a point pyramid transformer is investigated to model features with diverse resolutions or scales.
A multi-level transformer module is designed to aggregate contextual information from different levels of each scale and enhance their interactions.
arXiv Detail & Related papers (2021-04-28T08:39:21Z) - Cross-Level Cross-Scale Cross-Attention Network for Point Cloud
Representation [8.76786786874107]
Self-attention mechanism recently achieves impressive advancement in Natural Language Processing (NLP) and Image Processing domains.
We propose an end-to-end architecture, dubbed Cross-Level Cross-Scale Cross-Attention Network (CLCSCANet) for point cloud representation learning.
arXiv Detail & Related papers (2021-04-27T09:01:14Z) - Dual Transformer for Point Cloud Analysis [2.160196691362033]
We present a novel point cloud representation learning architecture, named Dual Transformer Network (DTNet)
Specifically, by aggregating the well-designed point-wise and channel-wise multi-head self-attention models simultaneously, DPCT module can capture much richer contextual dependencies semantically from perspective of position and channel.
Extensive quantitative and qualitative experiments on publicly available benchmarks demonstrate the effectiveness of our proposed transformer framework for the tasks of 3D point cloud classification and segmentation, achieving highly competitive performance in comparison with the state-of-the-art approaches.
arXiv Detail & Related papers (2021-04-27T08:41:02Z) - Few-shot 3D Point Cloud Semantic Segmentation [138.80825169240302]
We propose a novel attention-aware multi-prototype transductive few-shot point cloud semantic segmentation method.
Our proposed method shows significant and consistent improvements compared to baselines in different few-shot point cloud semantic segmentation settings.
arXiv Detail & Related papers (2020-06-22T08:05:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.