Bidirectional Feature Globalization for Few-shot Semantic Segmentation
of 3D Point Cloud Scenes
- URL: http://arxiv.org/abs/2208.06671v2
- Date: Wed, 17 Aug 2022 06:33:40 GMT
- Title: Bidirectional Feature Globalization for Few-shot Semantic Segmentation
of 3D Point Cloud Scenes
- Authors: Yongqiang Mao, Zonghao Guo, Xiaonan Lu, Zhiqiang Yuan, Haowen Guo
- Abstract summary: We propose a bidirectional feature globalization (BFG) approach to embed global perception to local point features.
With prototype-to-point globalization (Pr2PoG), the global perception is embedded to local point features based on similarity weights from sparse prototypes to dense point features.
The sparse prototypes of each class embedded with global perception are summarized to a single prototype for few-shot 3D segmentation.
- Score: 1.8374319565577157
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot segmentation of point cloud remains a challenging task, as there is
no effective way to convert local point cloud information to global
representation, which hinders the generalization ability of point features. In
this study, we propose a bidirectional feature globalization (BFG) approach,
which leverages the similarity measurement between point features and prototype
vectors to embed global perception to local point features in a bidirectional
fashion. With point-to-prototype globalization (Po2PrG), BFG aggregates local
point features to prototypes according to similarity weights from dense point
features to sparse prototypes. With prototype-to-point globalization (Pr2PoG),
the global perception is embedded to local point features based on similarity
weights from sparse prototypes to dense point features. The sparse prototypes
of each class embedded with global perception are summarized to a single
prototype for few-shot 3D segmentation based on the metric learning framework.
Extensive experiments on S3DIS and ScanNet demonstrate that BFG significantly
outperforms the state-of-the-art methods.
Related papers
- PointCore: Efficient Unsupervised Point Cloud Anomaly Detector Using
Local-Global Features [18.32982981001087]
We propose an unsupervised point cloud anomaly detection framework based on joint local-global features, termed PointCore.
To be specific, PointCore only requires a single memory bank to store local (coordinate) and global (PointMAE) representations.
Experiments on Real3D-AD dataset demonstrate that PointCore achieves competitive inference time and the best performance in both detection and localization.
arXiv Detail & Related papers (2024-03-04T07:51:46Z) - APPT : Asymmetric Parallel Point Transformer for 3D Point Cloud
Understanding [20.87092793669536]
Transformer-based networks have achieved impressive performance in 3D point cloud understanding.
To tackle these problems, we propose Asymmetric Parallel Point Transformer (APPT)
APPT is able to capture features globally throughout the entire network while focusing on local-detailed features.
arXiv Detail & Related papers (2023-03-31T06:11:02Z) - SphereVLAD++: Attention-based and Signal-enhanced Viewpoint Invariant
Descriptor [6.326554177747699]
We develop SphereVLAD++, an attention-enhanced viewpoint invariant place recognition method.
We show that SphereVLAD++ outperforms all relative state-of-the-art 3D place recognition methods under small or even totally reversed viewpoint differences.
arXiv Detail & Related papers (2022-07-06T20:32:43Z) - SemAffiNet: Semantic-Affine Transformation for Point Cloud Segmentation [94.11915008006483]
We propose SemAffiNet for point cloud semantic segmentation.
We conduct extensive experiments on the ScanNetV2 and NYUv2 datasets.
arXiv Detail & Related papers (2022-05-26T17:00:23Z) - SASA: Semantics-Augmented Set Abstraction for Point-based 3D Object
Detection [78.90102636266276]
We propose a novel set abstraction method named Semantics-Augmented Set Abstraction (SASA)
Based on the estimated point-wise foreground scores, we then propose a semantics-guided point sampling algorithm to help retain more important foreground points during down-sampling.
In practice, SASA shows to be effective in identifying valuable points related to foreground objects and improving feature learning for point-based 3D detection.
arXiv Detail & Related papers (2022-01-06T08:54:47Z) - Conformer: Local Features Coupling Global Representations for Visual
Recognition [72.9550481476101]
We propose a hybrid network structure, termed Conformer, to take advantage of convolutional operations and self-attention mechanisms for enhanced representation learning.
Experiments show that Conformer, under the comparable parameter complexity, outperforms the visual transformer (DeiT-B) by 2.3% on ImageNet.
arXiv Detail & Related papers (2021-05-09T10:00:03Z) - 3D Object Detection with Pointformer [29.935891419574602]
We propose Pointformer, a Transformer backbone designed for 3D point clouds to learn features effectively.
A Local Transformer module is employed to model interactions among points in a local region, which learns context-dependent region features at an object level.
A Global Transformer is designed to learn context-aware representations at the scene level.
arXiv Detail & Related papers (2020-12-21T15:12:54Z) - DH3D: Deep Hierarchical 3D Descriptors for Robust Large-Scale 6DoF
Relocalization [56.15308829924527]
We propose a Siamese network that jointly learns 3D local feature detection and description directly from raw 3D points.
For detecting 3D keypoints we predict the discriminativeness of the local descriptors in an unsupervised manner.
Experiments on various benchmarks demonstrate that our method achieves competitive results for both global point cloud retrieval and local point cloud registration.
arXiv Detail & Related papers (2020-07-17T20:21:22Z) - Global-Local Bidirectional Reasoning for Unsupervised Representation
Learning of 3D Point Clouds [109.0016923028653]
We learn point cloud representation by bidirectional reasoning between the local structures and the global shape without human supervision.
We show that our unsupervised model surpasses the state-of-the-art supervised methods on both synthetic and real-world 3D object classification datasets.
arXiv Detail & Related papers (2020-03-29T08:26:08Z) - A Rotation-Invariant Framework for Deep Point Cloud Analysis [132.91915346157018]
We introduce a new low-level purely rotation-invariant representation to replace common 3D Cartesian coordinates as the network inputs.
Also, we present a network architecture to embed these representations into features, encoding local relations between points and their neighbors, and the global shape structure.
We evaluate our method on multiple point cloud analysis tasks, including shape classification, part segmentation, and shape retrieval.
arXiv Detail & Related papers (2020-03-16T14:04:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.