ForestLPR: LiDAR Place Recognition in Forests Attentioning Multiple BEV Density Images
- URL: http://arxiv.org/abs/2503.04475v1
- Date: Thu, 06 Mar 2025 14:24:22 GMT
- Title: ForestLPR: LiDAR Place Recognition in Forests Attentioning Multiple BEV Density Images
- Authors: Yanqing Shen, Turcan Tuna, Marco Hutter, Cesar Cadena, Nanning Zheng,
- Abstract summary: We propose a robust LiDAR-based place recognition method for natural forests, ForestLPR.<n>Cross-sectional images of the forest's geometry at different heights contain the information needed to recognize revisiting a place.<n>Our approach utilizes a visual transformer as the shared backbone to produce sets of local descriptors.
- Score: 38.727720300337296
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Place recognition is essential to maintain global consistency in large-scale localization systems. While research in urban environments has progressed significantly using LiDARs or cameras, applications in natural forest-like environments remain largely under-explored. Furthermore, forests present particular challenges due to high self-similarity and substantial variations in vegetation growth over time. In this work, we propose a robust LiDAR-based place recognition method for natural forests, ForestLPR. We hypothesize that a set of cross-sectional images of the forest's geometry at different heights contains the information needed to recognize revisiting a place. The cross-sectional images are represented by \ac{bev} density images of horizontal slices of the point cloud at different heights. Our approach utilizes a visual transformer as the shared backbone to produce sets of local descriptors and introduces a multi-BEV interaction module to attend to information at different heights adaptively. It is followed by an aggregation layer that produces a rotation-invariant place descriptor. We evaluated the efficacy of our method extensively on real-world data from public benchmarks as well as robotic datasets and compared it against the state-of-the-art (SOTA) methods. The results indicate that ForestLPR has consistently good performance on all evaluations and achieves an average increase of 7.38\% and 9.11\% on Recall@1 over the closest competitor on intra-sequence loop closure detection and inter-sequence re-localization, respectively, validating our hypothesis
Related papers
- Zooming In on Fakes: A Novel Dataset for Localized AI-Generated Image Detection with Forgery Amplification Approach [69.01456182499486]
textbfBR-Gen is a large-scale dataset of 150,000 locally forged images with diverse scene-aware annotations.
textbfNFA-ViT is a Noise-guided Forgery Amplification Vision Transformer that enhances the detection of localized forgeries.
arXiv Detail & Related papers (2025-04-16T09:57:23Z) - Semantic segmentation of forest stands using deep learning [0.0]
Deep learning methods have demonstrated great potential in computer vision, yet their application to forest stand delineation remains unexplored.
This study presents a novel approach, framing stand delineation as a multiclass segmentation problem and applying a U-Net based DL framework.
The model was trained and evaluated using multispectral images, ALS data, and an existing stand map created by an expert interpreter.
arXiv Detail & Related papers (2025-04-03T10:47:25Z) - HOTFormerLoc: Hierarchical Octree Transformer for Versatile Lidar Place Recognition Across Ground and Aerial Views [30.77381516091565]
We present HOTFormerLoc, a novel and versatile Hierarchical Octree-based TransFormer, for large-scale 3D place recognition.
We propose an octree-based multi-scale attention mechanism that captures spatial and semantic features across granularities.
We introduce CS-Wild-Places, a novel 3D cross-source dataset featuring point cloud data from aerial and ground lidar scans captured in dense forests.
arXiv Detail & Related papers (2025-03-11T07:59:45Z) - NeRF-Accelerated Ecological Monitoring in Mixed-Evergreen Redwood Forest [0.0]
We present a comparison of MLS and NeRF forest reconstructions for the purpose of trunk diameter estimation in a mixed-evergreen Redwood forest.
We propose an improved DBH-estimation method using convex-hull modeling.
arXiv Detail & Related papers (2024-10-09T20:32:15Z) - Breaking the Frame: Visual Place Recognition by Overlap Prediction [53.17564423756082]
We propose a novel visual place recognition approach based on overlap prediction, called VOP.<n>VOP proceeds co-visible image sections by obtaining patch-level embeddings using a Vision Transformer backbone.<n>Our approach uses a voting mechanism to assess overlap scores for potential database images.
arXiv Detail & Related papers (2024-06-23T20:00:20Z) - BEVPlace: Learning LiDAR-based Place Recognition using Bird's Eye View
Images [20.30997801125592]
We explore the potential of a different representation in place recognition, i.e. bird's eye view (BEV) images.
A simple VGGNet trained on BEV images achieves comparable performance with the state-of-the-art place recognition methods in scenes of slight viewpoint changes.
We develop a method to estimate the position of the query cloud, extending the usage of place recognition.
arXiv Detail & Related papers (2023-02-28T05:37:45Z) - Spatial Likelihood Voting with Self-Knowledge Distillation for Weakly
Supervised Object Detection [54.24966006457756]
We propose a WSOD framework called the Spatial Likelihood Voting with Self-knowledge Distillation Network (SLV-SD Net)
SLV-SD Net converges region proposal localization without bounding box annotations.
Experiments on the PASCAL VOC 2007/2012 and MS-COCO datasets demonstrate the excellent performance of SLV-SD Net.
arXiv Detail & Related papers (2022-04-14T11:56:19Z) - PANet: Perspective-Aware Network with Dynamic Receptive Fields and
Self-Distilling Supervision for Crowd Counting [63.84828478688975]
We propose a novel perspective-aware approach called PANet to address the perspective problem.
Based on the observation that the size of the objects varies greatly in one image due to the perspective effect, we propose the dynamic receptive fields (DRF) framework.
The framework is able to adjust the receptive field by the dilated convolution parameters according to the input image, which helps the model to extract more discriminative features for each local region.
arXiv Detail & Related papers (2021-10-31T04:43:05Z) - LODE: Deep Local Deblurring and A New Benchmark [106.80269193462712]
We propose a novel framework, termed BLur-Aware DEblurring network (BladeNet)
BladeNet contains three components: the Local Blur Synthesis module generates locally blurred training pairs, the Local Blur Perception module automatically captures the locally blurred region and the Blur-guided Spatial Attention module guides the deblurring network with spatial attention.
We carry out extensive experiments on REDS and LODE datasets showing that BladeNet improves PSNR by 2.5dB over SotAs for local deblurring while keeping comparable performance for global deblurring.
arXiv Detail & Related papers (2021-09-19T15:57:28Z) - AdaZoom: Adaptive Zoom Network for Multi-Scale Object Detection in Large
Scenes [57.969186815591186]
Detection in large-scale scenes is a challenging problem due to small objects and extreme scale variation.
We propose a novel Adaptive Zoom (AdaZoom) network as a selective magnifier with flexible shape and focal length to adaptively zoom the focus regions for object detection.
arXiv Detail & Related papers (2021-06-19T03:30:22Z) - Instance segmentation of fallen trees in aerial color infrared imagery
using active multi-contour evolution with fully convolutional network-based
intensity priors [0.5276232626689566]
We introduce a framework for segmenting instances of a common object class by multiple active contour evolution over semantic segmentation maps of images.
We instantiate the proposed framework in the context of segmenting individual fallen stems from high-resolution aerial multispectral imagery.
arXiv Detail & Related papers (2021-05-05T11:54:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.