SUES-200: A Multi-height Multi-scene Cross-view Image Benchmark Across
Drone and Satellite
- URL: http://arxiv.org/abs/2204.10704v1
- Date: Fri, 22 Apr 2022 13:49:52 GMT
- Title: SUES-200: A Multi-height Multi-scene Cross-view Image Benchmark Across
Drone and Satellite
- Authors: Runzhe Zhu
- Abstract summary: The purpose of cross-view image matching is to match images acquired from different platforms of the same target scene.
SUES-200 is the first dataset that considers the differences generated by aerial photography of drones at different flight heights.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The purpose of cross-view image matching is to match images acquired from the
different platforms of the same target scene and then help positioning system
to infer the location of the target scene. With the rapid development of drone
technology, how to help Drone positioning or navigation through cross-view
matching technology has become a challenging research topic. However, the
accuracy of current cross-view matching models is still low, mainly because the
existing public datasets do not include the differences in images obtained by
drones at different heights, and the types of scenes are relatively
homogeneous, which makes the models unable to adapt to complex and changing
scenes. We propose a new cross-view dataset, SUES-200, to address these
issues.SUES-200 contains images acquired by the drone at four flight heights
and the corresponding satellite view images under the same target scene. To our
knowledge, SUES-200 is the first dataset that considers the differences
generated by aerial photography of drones at different flight heights. In
addition, we build a pipeline for efficient training testing and evaluation of
cross-view matching models. Then, we comprehensively evaluate the performance
of feature extractors with different CNN architectures on SUES-200 through an
evaluation system for cross-view matching models and propose a robust baseline
model. The experimental results show that SUES-200 can help the model learn
features with high discrimination at different heights. Evaluating indicators
of the matching system improves as the drone flight height gets higher because
the drone camera pose and the surrounding environment have less influence on
aerial photography.
Related papers
- Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - Towards Viewpoint Robustness in Bird's Eye View Segmentation [85.99907496019972]
We study how AV perception models are affected by changes in camera viewpoint.
Small changes to pitch, yaw, depth, or height of the camera at inference time lead to large drops in performance.
We introduce a technique for novel view synthesis and use it to transform collected data to the viewpoint of target rigs.
arXiv Detail & Related papers (2023-09-11T02:10:07Z) - VDD: Varied Drone Dataset for Semantic Segmentation [9.581655974280217]
We release a large-scale, densely labeled collection of 400 high-resolution images spanning 7 classes.
This dataset features various scenes in urban, industrial, rural, and natural areas, captured from different camera angles and under diverse lighting conditions.
We train seven state-of-the-art models on drone datasets as baselines.
arXiv Detail & Related papers (2023-05-23T02:16:14Z) - AirTrack: Onboard Deep Learning Framework for Long-Range Aircraft
Detection and Tracking [3.3773749296727535]
AirTrack is a real-time vision-only detect and tracking framework that respects the size, weight, and power constraints of sUAS systems.
We show that AirTrack outperforms state-of-the art baselines on the Amazon Airborne Object Tracking (AOT)
Empirical evaluations show that our system has a probability of track of more than 95% up to a range of 700m.
arXiv Detail & Related papers (2022-09-26T16:58:00Z) - Uncertainty Aware Multitask Pyramid Vision Transformer For UAV-Based
Object Re-Identification [38.19907319079833]
We propose a multitask learning approach, which employs a new multiscale architecture without convolution, Pyramid Vision Transformer (PVT) as the backbone for UAV-based object ReID.
By uncertainty modeling of intraclass variations, our proposed model can be jointly optimized using both uncertainty-aware object ID and camera ID information.
arXiv Detail & Related papers (2022-09-19T00:27:07Z) - Towards Robust Drone Vision in the Wild [1.14219428942199]
In this thesis, we propose the first image super-resolution dataset for drone vision.
We collect data at different altitudes and then propose pre-processing steps to align image pairs.
Two methods are proposed to build a robust image super-resolution network at different altitudes.
arXiv Detail & Related papers (2022-08-21T18:19:19Z) - VPAIR -- Aerial Visual Place Recognition and Localization in Large-scale
Outdoor Environments [49.82314641876602]
We present a new dataset named VPAIR.
The dataset was recorded on board a light aircraft flying at an altitude of more than 300 meters above ground.
The dataset covers a more than one hundred kilometers long trajectory over various types of challenging landscapes.
arXiv Detail & Related papers (2022-05-23T18:50:08Z) - Multi-Drone based Single Object Tracking with Agent Sharing Network [74.8198920355117]
Multi-Drone single Object Tracking dataset consists of 92 groups of video clips with 113,918 high resolution frames taken by two drones and 63 groups of video clips with 145,875 high resolution frames taken by three drones.
Agent sharing network (ASNet) is proposed by self-supervised template sharing and view-aware fusion of the target from multiple drones.
arXiv Detail & Related papers (2020-03-16T03:27:04Z) - University-1652: A Multi-view Multi-source Benchmark for Drone-based
Geo-localization [87.74121935246937]
We introduce a new multi-view benchmark for drone-based geo-localization, named University-1652.
University-1652 contains data from three platforms, i.e., synthetic drones, satellites and ground cameras of 1,652 university buildings around the world.
Experiments show that University-1652 helps the model to learn the viewpoint-invariant features and also has good generalization ability in the real-world scenario.
arXiv Detail & Related papers (2020-02-27T15:24:15Z) - Detection and Tracking Meet Drones Challenge [131.31749447313197]
This paper presents a review of object detection and tracking datasets and benchmarks, and discusses the challenges of collecting large-scale drone-based object detection and tracking datasets with manual annotations.
We describe our VisDrone dataset, which is captured over various urban/suburban areas of 14 different cities across China from North to South.
We provide a detailed analysis of the current state of the field of large-scale object detection and tracking on drones, and conclude the challenge as well as propose future directions.
arXiv Detail & Related papers (2020-01-16T00:11:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.