Benchmarking tree species classification from proximally-sensed laser scanning data: introducing the FOR-species20K dataset
- URL: http://arxiv.org/abs/2408.06507v1
- Date: Mon, 12 Aug 2024 21:47:15 GMT
- Title: Benchmarking tree species classification from proximally-sensed laser scanning data: introducing the FOR-species20K dataset
- Authors: Stefano Puliti, Emily R. Lines, Jana Müllerová, Julian Frey, Zoe Schindler, Adrian Straker, Matthew J. Allen, Lukas Winiwarter, Nataliia Rehush, Hristina Hristova, Brent Murray, Kim Calders, Louise Terryn, Nicholas Coops, Bernhard Höfle, Samuli Junttila, Martin Krůček, Grzegorz Krok, Kamil Král, Shaun R. Levick, Linda Luck, Azim Missarov, Martin Mokroš, Harry J. F. Owen, Krzysztof Stereńczak, Timo P. Pitkänen, Nicola Puletti, Ninni Saarinen, Chris Hopkinson, Chiara Torresan, Enrico Tomelleri, Hannah Weiser, Rasmus Astrup,
- Abstract summary: FOR-species20K benchmark was created, comprising over 20,000 tree point clouds from 33 species.
This dataset enables the benchmarking of DL models for tree species classification.
The top model, DetailView, was particularly robust, handling data imbalances well and generalizing effectively across tree sizes.
- Score: 1.2771525473423657
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Proximally-sensed laser scanning offers significant potential for automated forest data capture, but challenges remain in automatically identifying tree species without additional ground data. Deep learning (DL) shows promise for automation, yet progress is slowed by the lack of large, diverse, openly available labeled datasets of single tree point clouds. This has impacted the robustness of DL models and the ability to establish best practices for species classification. To overcome these challenges, the FOR-species20K benchmark dataset was created, comprising over 20,000 tree point clouds from 33 species, captured using terrestrial (TLS), mobile (MLS), and drone laser scanning (ULS) across various European forests, with some data from other regions. This dataset enables the benchmarking of DL models for tree species classification, including both point cloud-based (PointNet++, MinkNet, MLP-Mixer, DGCNNs) and multi-view image-based methods (SimpleView, DetailView, YOLOv5). 2D image-based models generally performed better (average OA = 0.77) than 3D point cloud-based models (average OA = 0.72), with consistent results across different scanning platforms and sensors. The top model, DetailView, was particularly robust, handling data imbalances well and generalizing effectively across tree sizes. The FOR-species20K dataset, available at https://zenodo.org/records/13255198, is a key resource for developing and benchmarking DL models for tree species classification using laser scanning data, providing a foundation for future advancements in the field.
Related papers
- PureForest: A Large-Scale Aerial Lidar and Aerial Imagery Dataset for Tree Species Classification in Monospecific Forests [0.0]
We present the PureForest dataset: a large-scale, open, multimodal dataset designed for tree species classification.
Most current public Lidar datasets for tree species classification have low diversity as they only span a small area of a few dozen annotated hectares at most.
In contrast, PureForest has 18 tree species grouped into 13 semantic classes, and spans 339 km$2$ across 449 distinct monospecific forests.
arXiv Detail & Related papers (2024-04-18T10:23:10Z) - SegmentAnyTree: A sensor and platform agnostic deep learning model for
tree segmentation using laser scanning data [15.438892555484616]
This research advances individual tree crown (ITC) segmentation in lidar data, using a deep learning model applicable to various laser scanning types.
It addresses the challenge of transferability across different data characteristics in 3D forest scene analysis.
The model, based on PointGroup architecture, is a 3D CNN with separate heads for semantic and instance segmentation.
arXiv Detail & Related papers (2024-01-28T19:47:17Z) - Leveraging Large-Scale Pretrained Vision Foundation Models for
Label-Efficient 3D Point Cloud Segmentation [67.07112533415116]
We present a novel framework that adapts various foundational models for the 3D point cloud segmentation task.
Our approach involves making initial predictions of 2D semantic masks using different large vision models.
To generate robust 3D semantic pseudo labels, we introduce a semantic label fusion strategy that effectively combines all the results via voting.
arXiv Detail & Related papers (2023-11-03T15:41:15Z) - AutoSynth: Learning to Generate 3D Training Data for Object Point Cloud
Registration [69.21282992341007]
Auto Synth automatically generates 3D training data for point cloud registration.
We replace the point cloud registration network with a much smaller surrogate network, leading to a $4056.43$ speedup.
Our results on TUD-L, LINEMOD and Occluded-LINEMOD evidence that a neural network trained on our searched dataset yields consistently better performance than the same one trained on the widely used ModelNet40 dataset.
arXiv Detail & Related papers (2023-09-20T09:29:44Z) - FOR-instance: a UAV laser scanning benchmark dataset for semantic and
instance segmentation of individual trees [0.06597195879147556]
FOR-instance dataset comprises five curated and ML-ready UAV-based laser scanning data collections.
The dataset is divided into development and test subsets, enabling method advancement and evaluation.
The inclusion of diameter at breast height data expands its utility to the measurement of a classic tree variable.
arXiv Detail & Related papers (2023-09-03T22:08:29Z) - Classification of structural building damage grades from multi-temporal
photogrammetric point clouds using a machine learning model trained on
virtual laser scanning data [58.720142291102135]
We present a novel approach to automatically assess multi-class building damage from real-world point clouds.
We use a machine learning model trained on virtual laser scanning (VLS) data.
The model yields high multi-target classification accuracies (overall accuracy: 92.0% - 95.1%)
arXiv Detail & Related papers (2023-02-24T12:04:46Z) - Learning 3D Human Pose Estimation from Dozens of Datasets using a
Geometry-Aware Autoencoder to Bridge Between Skeleton Formats [80.12253291709673]
We propose a novel affine-combining autoencoder (ACAE) method to perform dimensionality reduction on the number of landmarks.
Our approach scales to an extreme multi-dataset regime, where we use 28 3D human pose datasets to supervise one model.
arXiv Detail & Related papers (2022-12-29T22:22:49Z) - Tree Detection and Diameter Estimation Based on Deep Learning [0.0]
Tree perception is an essential building block toward autonomous forestry operations.
Deep neural network models trained on our datasets achieve a precision of 90.4% for tree detection.
Results offer promising avenues toward autonomous tree felling operations.
arXiv Detail & Related papers (2022-10-31T15:51:32Z) - Omnidata: A Scalable Pipeline for Making Multi-Task Mid-Level Vision
Datasets from 3D Scans [103.92680099373567]
This paper introduces a pipeline to parametrically sample and render multi-task vision datasets from comprehensive 3D scans from the real world.
Changing the sampling parameters allows one to "steer" the generated datasets to emphasize specific information.
Common architectures trained on a generated starter dataset reached state-of-the-art performance on multiple common vision tasks and benchmarks.
arXiv Detail & Related papers (2021-10-11T04:21:46Z) - Campus3D: A Photogrammetry Point Cloud Benchmark for Hierarchical
Understanding of Outdoor Scene [76.4183572058063]
We present a richly-annotated 3D point cloud dataset for multiple outdoor scene understanding tasks.
The dataset has been point-wisely annotated with both hierarchical and instance-based labels.
We formulate a hierarchical learning problem for 3D point cloud segmentation and propose a measurement evaluating consistency across various hierarchies.
arXiv Detail & Related papers (2020-08-11T19:10:32Z) - Tree Annotations in LiDAR Data Using Point Densities and Convolutional
Neural Networks [9.374986160570034]
We present three automatic methods for annotating trees in LiDAR data.
The first method requires high density point clouds and uses certain LiDAR data attributes for the purpose of tree identification, achieving almost 90% accuracy.
The second method uses a voxel-based 3D Convolutional Neural Network on low density LiDAR datasets and is able to identify most large trees accurately but struggles with smaller ones due to the voxelisation process.
The third method is a scaled version of the PointNet++ method and works directly on outdoor point clouds and achieves an F_score of 82.1% on the ISPRS benchmark dataset
arXiv Detail & Related papers (2020-06-09T23:50:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.