CGS-Net: Aggregating Colour, Geometry and Semantic Features for
Large-Scale Indoor Place Recognition
- URL: http://arxiv.org/abs/2202.02070v1
- Date: Fri, 4 Feb 2022 10:51:25 GMT
- Title: CGS-Net: Aggregating Colour, Geometry and Semantic Features for
Large-Scale Indoor Place Recognition
- Authors: Yuhang Ming, Xingrui Yang, Guofeng Zhang, Andrew Calway
- Abstract summary: We describe an approach to large-scale indoor place recognition that aggregates low-level colour and geometric features with high-level semantic features.
We use a deep learning network that takes in RGB point clouds and extracts local features with five 3-D kernel point convolutional layers.
We specifically train the KPConv layers on the semantic segmentation task to ensure that the extracted local features are semantically meaningful.
- Score: 6.156387608994791
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We describe an approach to large-scale indoor place recognition that
aggregates low-level colour and geometric features with high-level semantic
features. We use a deep learning network that takes in RGB point clouds and
extracts local features with five 3-D kernel point convolutional (KPConv)
layers. We specifically train the KPConv layers on the semantic segmentation
task to ensure that the extracted local features are semantically meaningful.
Then, feature maps from all the five KPConv layers are concatenated together
and fed into the NetVLAD layer to generate the global descriptors. The approach
is trained and evaluated using a large-scale indoor place recognition dataset
derived from the ScanNet dataset, with a test set comprising 3,608 point clouds
generated from 100 different rooms. Comparison with a traditional feature based
method and three state-of-the-art deep learning methods demonstrate that the
approach significantly outperforms all four methods, achieving, for example, a
top-3 average recall rate of 75% compared with 41% for the closest rival
method.
Related papers
- Continual Learning in 3D Point Clouds: Employing Spectral Techniques for Exemplar Selection [0.40964539027092906]
We introduce a novel framework for Continual Learning in 3D object classification (CL3D)
Our approach is based on the selection of prototypes from each class using spectral clustering.
We conduct experiments on the ModelNet40, ShapeNet, and ScanNet datasets, achieving state-of-the-art accuracy exclusively through the use of input space features.
arXiv Detail & Related papers (2024-09-12T20:34:34Z) - AEGIS-Net: Attention-guided Multi-Level Feature Aggregation for Indoor
Place Recognition [12.728087388529028]
AEGIS-Net is a novel indoor place recognition model that takes in RGB point clouds and generates global place descriptors.
Our AEGIS-Net is made of a semantic encoder, a semantic decoder and an attention-guided feature embedding.
We evaluate our AEGIS-Net on the ScanNetPR dataset and compare its performance with a pre-deep-learning feature-based method and five state-of-the-art deep-learning-based methods.
arXiv Detail & Related papers (2023-12-15T05:09:08Z) - Dynamic Clustering Transformer Network for Point Cloud Segmentation [23.149220817575195]
We propose a novel 3D point cloud representation network, called Dynamic Clustering Transformer Network (DCTNet)
It has an encoder-decoder architecture, allowing for both local and global feature learning.
Our method was evaluated on an object-based dataset (ShapeNet), an urban navigation dataset (Toronto-3D), and a multispectral LiDAR dataset.
arXiv Detail & Related papers (2023-05-30T01:11:05Z) - SemAffiNet: Semantic-Affine Transformation for Point Cloud Segmentation [94.11915008006483]
We propose SemAffiNet for point cloud semantic segmentation.
We conduct extensive experiments on the ScanNetV2 and NYUv2 datasets.
arXiv Detail & Related papers (2022-05-26T17:00:23Z) - Two Heads are Better than One: Geometric-Latent Attention for Point
Cloud Classification and Segmentation [10.2254921311882]
We present an innovative two-headed attention layer that combines geometric and latent features to segment a 3D scene into meaningful subsets.
Each head combines local and global information, using either the geometric or latent features, of a neighborhood of points and uses this information to learn better local relationships.
arXiv Detail & Related papers (2021-10-30T11:20:56Z) - Learning Semantic Segmentation of Large-Scale Point Clouds with Random
Sampling [52.464516118826765]
We introduce RandLA-Net, an efficient and lightweight neural architecture to infer per-point semantics for large-scale point clouds.
The key to our approach is to use random point sampling instead of more complex point selection approaches.
Our RandLA-Net can process 1 million points in a single pass up to 200x faster than existing approaches.
arXiv Detail & Related papers (2021-07-06T05:08:34Z) - A Holistically-Guided Decoder for Deep Representation Learning with
Applications to Semantic Segmentation and Object Detection [74.88284082187462]
One common strategy is to adopt dilated convolutions in the backbone networks to extract high-resolution feature maps.
We propose one novel holistically-guided decoder which is introduced to obtain the high-resolution semantic-rich feature maps.
arXiv Detail & Related papers (2020-12-18T10:51:49Z) - Campus3D: A Photogrammetry Point Cloud Benchmark for Hierarchical
Understanding of Outdoor Scene [76.4183572058063]
We present a richly-annotated 3D point cloud dataset for multiple outdoor scene understanding tasks.
The dataset has been point-wisely annotated with both hierarchical and instance-based labels.
We formulate a hierarchical learning problem for 3D point cloud segmentation and propose a measurement evaluating consistency across various hierarchies.
arXiv Detail & Related papers (2020-08-11T19:10:32Z) - DH3D: Deep Hierarchical 3D Descriptors for Robust Large-Scale 6DoF
Relocalization [56.15308829924527]
We propose a Siamese network that jointly learns 3D local feature detection and description directly from raw 3D points.
For detecting 3D keypoints we predict the discriminativeness of the local descriptors in an unsupervised manner.
Experiments on various benchmarks demonstrate that our method achieves competitive results for both global point cloud retrieval and local point cloud registration.
arXiv Detail & Related papers (2020-07-17T20:21:22Z) - PointGroup: Dual-Set Point Grouping for 3D Instance Segmentation [111.7241018610573]
We present PointGroup, a new end-to-end bottom-up architecture for instance segmentation.
We design a two-branch network to extract point features and predict semantic labels and offsets, for shifting each point towards its respective instance centroid.
A clustering component is followed to utilize both the original and offset-shifted point coordinate sets, taking advantage of their complementary strength.
We conduct extensive experiments on two challenging datasets, ScanNet v2 and S3DIS, on which our method achieves the highest performance, 63.6% and 64.0%, compared to 54.9% and 54.4% achieved by former best
arXiv Detail & Related papers (2020-04-03T16:26:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.