AttDLNet: Attention-based DL Network for 3D LiDAR Place Recognition
- URL: http://arxiv.org/abs/2106.09637v1
- Date: Thu, 17 Jun 2021 16:34:37 GMT
- Title: AttDLNet: Attention-based DL Network for 3D LiDAR Place Recognition
- Authors: Tiago Barros, Lu\'is Garrote, Ricardo Pereira, Cristiano Premebida,
Urbano J. Nunes
- Abstract summary: This paper proposes a novel 3D LiDAR-based deep learning network named AttDLNet.
It exploits an attention mechanism to selectively focus on long-range context and interfeature relationships.
Results show that the encoder network features are already very descriptive, but adding attention to the network further improves performance.
- Score: 0.6352264764099531
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep networks have been progressively adapted to new sensor modalities,
namely to 3D LiDAR, which led to unprecedented achievements in autonomous
vehicle-related applications such as place recognition. One of the main
challenges of deep models in place recognition is to extract efficient and
descriptive feature representations that relate places based on their
similarity. To address the problem of place recognition using LiDAR data, this
paper proposes a novel 3D LiDAR-based deep learning network (named AttDLNet)
that comprises an encoder network and exploits an attention mechanism to
selectively focus on long-range context and interfeature relationships. The
proposed network is trained and validated on the KITTI dataset, using the
cosine loss for training and a retrieval-based place recognition pipeline for
validation. Additionally, an ablation study is presented to assess the best
network configuration. Results show that the encoder network features are
already very descriptive, but adding attention to the network further improves
performance. From the ablation study, results indicate that the middle encoder
layers have the highest mean performance, while deeper layers are more robust
to orientation change. The code is publicly available on the project website:
https://github.com/Cybonic/ AttDLNet
Related papers
- Densely Decoded Networks with Adaptive Deep Supervision for Medical
Image Segmentation [19.302294715542175]
We propose densely decoded networks (ddn), by selectively introducing 'crutch' network connections.
Such 'crutch' connections in each upsampling stage of the network decoder enhance target localization.
We also present a training strategy based on adaptive deep supervision (ads), which exploits and adapts specific attributes of input dataset.
arXiv Detail & Related papers (2024-02-05T00:44:57Z) - SERNet-Former: Semantic Segmentation by Efficient Residual Network with Attention-Boosting Gates and Attention-Fusion Networks [0.0]
This research proposes an encoder-decoder architecture with a unique efficient residual network, Efficient-ResNet.
Attention-boosting gates (AbGs) and attention-boosting modules (AbMs) are deployed by aiming to fuse the equivariant and feature-based semantic information with the equivalent sizes of the output of global context.
Our network is tested on the challenging CamVid and Cityscapes datasets, and the proposed methods reveal significant improvements on the residual networks.
arXiv Detail & Related papers (2024-01-28T19:58:19Z) - Active search and coverage using point-cloud reinforcement learning [50.741409008225766]
This paper presents an end-to-end deep reinforcement learning solution for target search and coverage.
We show that deep hierarchical feature learning works for RL and that by using farthest point sampling (FPS) we can reduce the amount of points.
We also show that multi-head attention for point-clouds helps to learn the agent faster but converges to the same outcome.
arXiv Detail & Related papers (2023-12-18T18:16:30Z) - HEDNet: A Hierarchical Encoder-Decoder Network for 3D Object Detection
in Point Clouds [19.1921315424192]
3D object detection in point clouds is important for autonomous driving systems.
A primary challenge in 3D object detection stems from the sparse distribution of points within the 3D scene.
We propose HEDNet, a hierarchical encoder-decoder network for 3D object detection.
arXiv Detail & Related papers (2023-10-31T07:32:08Z) - NAF: Neural Attenuation Fields for Sparse-View CBCT Reconstruction [79.13750275141139]
This paper proposes a novel and fast self-supervised solution for sparse-view CBCT reconstruction.
The desired attenuation coefficients are represented as a continuous function of 3D spatial coordinates, parameterized by a fully-connected deep neural network.
A learning-based encoder entailing hash coding is adopted to help the network capture high-frequency details.
arXiv Detail & Related papers (2022-09-29T04:06:00Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - Paint and Distill: Boosting 3D Object Detection with Semantic Passing
Network [70.53093934205057]
3D object detection task from lidar or camera sensors is essential for autonomous driving.
We propose a novel semantic passing framework, named SPNet, to boost the performance of existing lidar-based 3D detection models.
arXiv Detail & Related papers (2022-07-12T12:35:34Z) - Simplified Learning of CAD Features Leveraging a Deep Residual
Autoencoder [3.567248644184455]
In computer vision, deep residual neural networks like EfficientNet have set new standards in terms of robustness and accuracy.
One key problem underlying the training of deep neural networks is the immanent lack of a sufficient amount of training data.
We present a deep residual 3D autoencoder based on the EfficientNet architecture, intended for transfer learning tasks related to 3D CAD model assessment.
arXiv Detail & Related papers (2022-02-21T10:27:55Z) - Location-Sensitive Visual Recognition with Cross-IOU Loss [177.86369890708457]
This paper proposes a unified solution named location-sensitive network (LSNet) for object detection, instance segmentation, and pose estimation.
Based on a deep neural network as the backbone, LSNet predicts an anchor point and a set of landmarks which together define the shape of the target object.
arXiv Detail & Related papers (2021-04-11T02:17:14Z) - PLADE-Net: Towards Pixel-Level Accuracy for Self-Supervised Single-View
Depth Estimation with Neural Positional Encoding and Distilled Matting Loss [49.66736599668501]
We propose a self-supervised single-view pixel-level accurate depth estimation network, called PLADE-Net.
Our method shows unprecedented accuracy levels, exceeding 95% in terms of the $delta1$ metric on the KITTI dataset.
arXiv Detail & Related papers (2021-03-12T15:54:46Z) - Attentional Local Contrast Networks for Infrared Small Target Detection [15.882749652217653]
We propose a novel model-driven deep network for infrared small target detection.
We modularize a conventional local contrast measure method as a depth-wise parameterless nonlinear feature refinement layer in an end-to-end network.
We conduct detailed ablation studies with varying network depths to empirically verify the effectiveness and efficiency of each component in our network architecture.
arXiv Detail & Related papers (2020-12-15T19:33:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.