Skyline variations allow estimating distance to trees on landscape
photos using semantic segmentation
- URL: http://arxiv.org/abs/2201.08816v1
- Date: Fri, 14 Jan 2022 12:31:02 GMT
- Title: Skyline variations allow estimating distance to trees on landscape
photos using semantic segmentation
- Authors: Laura Martinez-Sanchez, Daniele Borio, Rapha\"el d'Andrimont, Marijn
van der Velde
- Abstract summary: We show that variations in the skyline of landscape photos can be used to estimate distances to trees on the horizon.
A methodology based on the variations of the skyline has been developed and used to investigate potential relationships with the distance to skyline objects.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Approximate distance estimation can be used to determine fundamental
landscape properties including complexity and openness. We show that variations
in the skyline of landscape photos can be used to estimate distances to trees
on the horizon. A methodology based on the variations of the skyline has been
developed and used to investigate potential relationships with the distance to
skyline objects. The skyline signal, defined by the skyline height expressed in
pixels, was extracted for several Land Use/Cover Area frame Survey (LUCAS)
landscape photos. Photos were semantically segmented with DeepLabV3+ trained
with the Common Objects in Context (COCO) dataset. This provided pixel-level
classification of the objects forming the skyline. A Conditional Random Fields
(CRF) algorithm was also applied to increase the details of the skyline signal.
Three metrics, able to capture the skyline signal variations, were then
considered for the analysis. These metrics shows a functional relationship with
distance for the class of trees, whose contours have a fractal nature. In
particular, regression analysis was performed against 475 ortho-photo based
distance measurements, and, in the best case, a R2 score equal to 0.47 was
achieved. This is an encouraging result which shows the potential of skyline
variation metrics for inferring distance related information.
Related papers
- OAM-TCD: A globally diverse dataset of high-resolution tree cover maps [8.336960607169175]
We present a novel open-access dataset for individual tree crown delineation (TCD) in high-resolution aerial imagery sourced from OpenMap (OAM)
Our dataset, OAM-TCD, comprises 5072 2048x2048px images at 10 cm/px resolution with associated human-labeled instance masks for over 280k individual and 56k groups of trees.
Using our dataset, we train reference instance and semantic segmentation models that compare favorably to existing state-of-the-art models.
arXiv Detail & Related papers (2024-07-16T14:11:29Z) - Learning Dense Flow Field for Highly-accurate Cross-view Camera
Localization [15.89357790711828]
This paper addresses the problem of estimating the 3-DoF camera pose for a ground-level image with respect to a satellite image.
We propose a novel end-to-end approach that leverages the learning of dense pixel-wise flow fields in pairs of ground and satellite images.
Our approach reduces the median localization error by 89%, 19%, 80% and 35% on the KITTI, Ford multi-AV, VIGOR and Oxford RobotCar datasets.
arXiv Detail & Related papers (2023-09-27T10:26:26Z) - Linear features segmentation from aerial images [0.0]
We present a method for classifying and segmenting city road traffic dashed lines from aerial images using deep learning models such as U-Net and SegNet.
The annotated data is used to train these models, which are then used to classify and segment the aerial image into two classes: dashed lines and non-dashed lines.
We also extracted the x and y coordinates of each dashed line from the segmentation output, which can be used by city planners to construct a CAD file for digital visualization of the roads.
arXiv Detail & Related papers (2022-12-23T18:51:14Z) - SensatUrban: Learning Semantics from Urban-Scale Photogrammetric Point
Clouds [52.624157840253204]
We introduce SensatUrban, an urban-scale UAV photogrammetry point cloud dataset consisting of nearly three billion points collected from three UK cities, covering 7.6 km2.
Each point in the dataset has been labelled with fine-grained semantic annotations, resulting in a dataset that is three times the size of the previous existing largest photogrammetric point cloud dataset.
arXiv Detail & Related papers (2022-01-12T14:48:11Z) - Semantic Segmentation on Swiss3DCities: A Benchmark Study on Aerial
Photogrammetric 3D Pointcloud Dataset [67.44497676652173]
We introduce a new outdoor urban 3D pointcloud dataset, covering a total area of 2.7 $km2$, sampled from three Swiss cities.
The dataset is manually annotated for semantic segmentation with per-point labels, and is built using photogrammetry from images acquired by multirotors equipped with high-resolution cameras.
arXiv Detail & Related papers (2020-12-23T21:48:47Z) - Baseline and Triangulation Geometry in a Standard Plenoptic Camera [6.719751155411075]
We present a geometrical light field model allowing triangulation to be applied to a plenoptic camera.
It is shown that distance estimates from our novel method match those of real objects placed in front of the camera.
arXiv Detail & Related papers (2020-10-09T15:31:14Z) - Campus3D: A Photogrammetry Point Cloud Benchmark for Hierarchical
Understanding of Outdoor Scene [76.4183572058063]
We present a richly-annotated 3D point cloud dataset for multiple outdoor scene understanding tasks.
The dataset has been point-wisely annotated with both hierarchical and instance-based labels.
We formulate a hierarchical learning problem for 3D point cloud segmentation and propose a measurement evaluating consistency across various hierarchies.
arXiv Detail & Related papers (2020-08-11T19:10:32Z) - Single View Metrology in the Wild [94.7005246862618]
We present a novel approach to single view metrology that can recover the absolute scale of a scene represented by 3D heights of objects or camera height above the ground.
Our method relies on data-driven priors learned by a deep network specifically designed to imbibe weakly supervised constraints from the interplay of the unknown camera with 3D entities such as object heights.
We demonstrate state-of-the-art qualitative and quantitative results on several datasets as well as applications including virtual object insertion.
arXiv Detail & Related papers (2020-07-18T22:31:33Z) - Where am I looking at? Joint Location and Orientation Estimation by
Cross-View Matching [95.64702426906466]
Cross-view geo-localization is a problem given a large-scale database of geo-tagged aerial images.
Knowing orientation between ground and aerial images can significantly reduce matching ambiguity between these two views.
We design a Dynamic Similarity Matching network to estimate cross-view orientation alignment during localization.
arXiv Detail & Related papers (2020-05-08T05:21:16Z) - Deep 3D Capture: Geometry and Reflectance from Sparse Multi-View Images [59.906948203578544]
We introduce a novel learning-based method to reconstruct the high-quality geometry and complex, spatially-varying BRDF of an arbitrary object.
We first estimate per-view depth maps using a deep multi-view stereo network.
These depth maps are used to coarsely align the different views.
We propose a novel multi-view reflectance estimation network architecture.
arXiv Detail & Related papers (2020-03-27T21:28:54Z) - Evaluation of Cross-View Matching to Improve Ground Vehicle Localization
with Aerial Perception [17.349420462716886]
Cross-view matching refers to the problem of finding the closest match for a given query ground view image to one from a database of aerial images.
In this paper, we evaluate cross-view matching for the task of localizing a ground vehicle over a longer trajectory.
arXiv Detail & Related papers (2020-03-13T23:59:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.