UAVStereo: A Multiple Resolution Dataset for Stereo Matching in UAV
Scenarios
- URL: http://arxiv.org/abs/2302.10082v1
- Date: Mon, 20 Feb 2023 16:45:27 GMT
- Title: UAVStereo: A Multiple Resolution Dataset for Stereo Matching in UAV
Scenarios
- Authors: Zhang Xiaoyi, Cao Xuefeng, Yu Anzhu, Yu Wenshuai, Li Zhenqi, Quan
Yujun
- Abstract summary: This paper constructs a multi-resolution UAV scenario dataset, called UAVStereo, with over 34k stereo image pairs covering 3 typical scenes.
In this paper, we evaluate traditional and state-of-the-art deep learning methods, highlighting their limitations in addressing challenges in UAV scenarios.
- Score: 0.6524460254566905
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stereo matching is a fundamental task for 3D scene reconstruction. Recently,
deep learning based methods have proven effective on some benchmark datasets,
such as KITTI and Scene Flow. UAVs (Unmanned Aerial Vehicles) are commonly
utilized for surface observation, and their captured images are frequently used
for detailed 3D reconstruction due to high resolution and low-altitude
acquisition. At present, the mainstream supervised learning network requires a
significant amount of training data with ground-truth labels to learn model
parameters. However, due to the scarcity of UAV stereo matching datasets, the
learning-based network cannot be applied to UAV images. To facilitate further
research, this paper proposes a novel pipeline to generate accurate and dense
disparity maps using detailed meshes reconstructed by UAV images and LiDAR
point clouds. Through the proposed pipeline, this paper constructs a
multi-resolution UAV scenario dataset, called UAVStereo, with over 34k stereo
image pairs covering 3 typical scenes. As far as we know, UAVStereo is the
first stereo matching dataset of UAV low-altitude scenarios. The dataset
includes synthetic and real stereo pairs to enable generalization from the
synthetic domain to the real domain. Furthermore, our UAVStereo dataset
provides multi-resolution and multi-scene images pairs to accommodate a variety
of sensors and environments. In this paper, we evaluate traditional and
state-of-the-art deep learning methods, highlighting their limitations in
addressing challenges in UAV scenarios and offering suggestions for future
research. The dataset is available at
https://github.com/rebecca0011/UAVStereo.git
Related papers
- Game4Loc: A UAV Geo-Localization Benchmark from Game Data [0.0]
We introduce a more practical UAV geo-localization task including partial matches of cross-view paired data.
Experiments demonstrate the effectiveness of our data and training method for UAV geo-localization.
arXiv Detail & Related papers (2024-09-25T13:33:28Z) - MMScan: A Multi-Modal 3D Scene Dataset with Hierarchical Grounded Language Annotations [55.022519020409405]
This paper builds the first largest ever multi-modal 3D scene dataset and benchmark with hierarchical grounded language annotations, MMScan.
The resulting multi-modal 3D dataset encompasses 1.4M meta-annotated captions on 109k objects and 7.7k regions as well as over 3.04M diverse samples for 3D visual grounding and question-answering benchmarks.
arXiv Detail & Related papers (2024-06-13T17:59:30Z) - UAV-VisLoc: A Large-scale Dataset for UAV Visual Localization [20.37586403749362]
We present a large-scale dataset, UAV-VisLoc, to facilitate the UAV visual localization task.
Our dataset includes 6,742 drone images and 11 satellite maps, with metadata such as latitude, longitude, altitude, and capture date.
arXiv Detail & Related papers (2024-05-20T10:24:10Z) - UAVD4L: A Large-Scale Dataset for UAV 6-DoF Localization [14.87295056434887]
We introduce a large-scale 6-DoF UAV dataset for localization (UAVD4L)
We develop a two-stage 6-DoF localization pipeline (UAVLoc), which consists of offline synthetic data generation and online visual localization.
Results on the new dataset demonstrate the effectiveness of the proposed approach.
arXiv Detail & Related papers (2024-01-11T15:19:21Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - Investigation of UAV Detection in Images with Complex Backgrounds and
Rainy Artifacts [20.20609511526255]
Vision-based object detection methods have been developed for UAV detection.
UAV detection in images with complex backgrounds and weather artifacts like rain has yet to be reasonably studied.
This work also focuses on benchmarking state-of-the-art object detection models.
arXiv Detail & Related papers (2023-05-25T19:54:33Z) - Towards Scale Consistent Monocular Visual Odometry by Learning from the
Virtual World [83.36195426897768]
We propose VRVO, a novel framework for retrieving the absolute scale from virtual data.
We first train a scale-aware disparity network using both monocular real images and stereo virtual data.
The resulting scale-consistent disparities are then integrated with a direct VO system.
arXiv Detail & Related papers (2022-03-11T01:51:54Z) - Vision-Based UAV Self-Positioning in Low-Altitude Urban Environments [20.69412701553767]
Unmanned Aerial Vehicles (UAVs) rely on satellite systems for stable positioning.
In such situations, vision-based techniques can serve as an alternative, ensuring the self-positioning capability of UAVs.
This paper presents a new dataset, DenseUAV, which is the first publicly available dataset designed for the UAV self-positioning task.
arXiv Detail & Related papers (2022-01-23T07:18:55Z) - Self-Supervised Depth Completion for Active Stereo [55.79929735390945]
Active stereo systems are widely used in the robotics industry due to their low cost and high quality depth maps.
These depth sensors suffer from stereo artefacts and do not provide dense depth estimates.
We present the first self-supervised depth completion method for active stereo systems that predicts accurate dense depth maps.
arXiv Detail & Related papers (2021-10-07T07:33:52Z) - Self-supervised Audiovisual Representation Learning for Remote Sensing Data [96.23611272637943]
We propose a self-supervised approach for pre-training deep neural networks in remote sensing.
By exploiting the correspondence between geo-tagged audio recordings and remote sensing, this is done in a completely label-free manner.
We show that our approach outperforms existing pre-training strategies for remote sensing imagery.
arXiv Detail & Related papers (2021-08-02T07:50:50Z) - OpenREALM: Real-time Mapping for Unmanned Aerial Vehicles [62.997667081978825]
OpenREALM is a real-time mapping framework for Unmanned Aerial Vehicles (UAVs)
Different modes of operation allow OpenREALM to perform simple stitching assuming an approximate plane ground.
In all modes incremental progress of the resulting map can be viewed live by an operator on the ground.
arXiv Detail & Related papers (2020-09-22T12:28:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.