Forest Inspection Dataset for Aerial Semantic Segmentation and Depth
Estimation
- URL: http://arxiv.org/abs/2403.06621v1
- Date: Mon, 11 Mar 2024 11:26:44 GMT
- Title: Forest Inspection Dataset for Aerial Semantic Segmentation and Depth
Estimation
- Authors: Bianca-Cerasela-Zelia Blaga and Sergiu Nedevschi
- Abstract summary: We introduce a new large aerial dataset for forest inspection.
It contains both real-world and virtual recordings of natural environments.
We develop a framework to assess the deforestation degree of an area.
- Score: 6.635604919499181
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Humans use UAVs to monitor changes in forest environments since they are
lightweight and provide a large variety of surveillance data. However, their
information does not present enough details for understanding the scene which
is needed to assess the degree of deforestation. Deep learning algorithms must
be trained on large amounts of data to output accurate interpretations, but
ground truth recordings of annotated forest imagery are not available. To solve
this problem, we introduce a new large aerial dataset for forest inspection
which contains both real-world and virtual recordings of natural environments,
with densely annotated semantic segmentation labels and depth maps, taken in
different illumination conditions, at various altitudes and recording angles.
We test the performance of two multi-scale neural networks for solving the
semantic segmentation task (HRNet and PointFlow network), studying the impact
of the various acquisition conditions and the capabilities of transfer learning
from virtual to real data. Our results showcase that the best results are
obtained when the training is done on a dataset containing a large variety of
scenarios, rather than separating the data into specific categories. We also
develop a framework to assess the deforestation degree of an area.
Related papers
- Towards general deep-learning-based tree instance segmentation models [0.0]
Deep-learning methods have been proposed which show the potential of learning to segment trees.
We use seven diverse datasets found in literature to gain insights into the generalization capabilities under domain-shift.
Our results suggest that a generalization from coniferous dominated sparse point clouds to deciduous dominated high-resolution point clouds is possible.
arXiv Detail & Related papers (2024-05-03T12:42:43Z) - Training point-based deep learning networks for forest segmentation with synthetic data [0.0]
We develop a realistic simulator that procedurally generates synthetic forest scenes.
We conduct a comparative study of different state-of-the-art point-based deep learning networks for forest segmentation.
arXiv Detail & Related papers (2024-03-21T04:01:26Z) - Rethinking Transformers Pre-training for Multi-Spectral Satellite
Imagery [78.43828998065071]
Recent advances in unsupervised learning have demonstrated the ability of large vision models to achieve promising results on downstream tasks.
Such pre-training techniques have also been explored recently in the remote sensing domain due to the availability of large amount of unlabelled data.
In this paper, we re-visit transformers pre-training and leverage multi-scale information that is effectively utilized with multiple modalities.
arXiv Detail & Related papers (2024-03-08T16:18:04Z) - SegmentAnyTree: A sensor and platform agnostic deep learning model for
tree segmentation using laser scanning data [15.438892555484616]
This research advances individual tree crown (ITC) segmentation in lidar data, using a deep learning model applicable to various laser scanning types.
It addresses the challenge of transferability across different data characteristics in 3D forest scene analysis.
The model, based on PointGroup architecture, is a 3D CNN with separate heads for semantic and instance segmentation.
arXiv Detail & Related papers (2024-01-28T19:47:17Z) - FOR-instance: a UAV laser scanning benchmark dataset for semantic and
instance segmentation of individual trees [0.06597195879147556]
FOR-instance dataset comprises five curated and ML-ready UAV-based laser scanning data collections.
The dataset is divided into development and test subsets, enabling method advancement and evaluation.
The inclusion of diameter at breast height data expands its utility to the measurement of a classic tree variable.
arXiv Detail & Related papers (2023-09-03T22:08:29Z) - Do More With What You Have: Transferring Depth-Scale from Labeled to Unlabeled Domains [43.16293941978469]
Self-supervised depth estimators result in up-to-scale predictions that are linearly correlated to their absolute depth values across the domain.
We show that aligning the field-of-view of two datasets prior to training results in a common linear relationship for both domains.
We use this observed property to transfer the depth-scale from source datasets that have absolute depth labels to new target datasets that lack these measurements.
arXiv Detail & Related papers (2023-03-14T07:07:34Z) - Detection Hub: Unifying Object Detection Datasets via Query Adaptation
on Language Embedding [137.3719377780593]
A new design (named Detection Hub) is dataset-aware and category-aligned.
It mitigates the dataset inconsistency and provides coherent guidance for the detector to learn across multiple datasets.
The categories across datasets are semantically aligned into a unified space by replacing one-hot category representations with word embedding.
arXiv Detail & Related papers (2022-06-07T17:59:44Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - Self-supervised Audiovisual Representation Learning for Remote Sensing Data [96.23611272637943]
We propose a self-supervised approach for pre-training deep neural networks in remote sensing.
By exploiting the correspondence between geo-tagged audio recordings and remote sensing, this is done in a completely label-free manner.
We show that our approach outperforms existing pre-training strategies for remote sensing imagery.
arXiv Detail & Related papers (2021-08-02T07:50:50Z) - Speak2Label: Using Domain Knowledge for Creating a Large Scale Driver
Gaze Zone Estimation Dataset [55.391532084304494]
Driver Gaze in the Wild dataset contains 586 recordings, captured during different times of the day including evenings.
Driver Gaze in the Wild dataset contains 338 subjects with an age range of 18-63 years.
arXiv Detail & Related papers (2020-04-13T14:47:34Z) - Laplacian Denoising Autoencoder [114.21219514831343]
We propose to learn data representations with a novel type of denoising autoencoder.
The noisy input data is generated by corrupting latent clean data in the gradient domain.
Experiments on several visual benchmarks demonstrate that better representations can be learned with the proposed approach.
arXiv Detail & Related papers (2020-03-30T16:52:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.