University-1652: A Multi-view Multi-source Benchmark for Drone-based
Geo-localization
- URL: http://arxiv.org/abs/2002.12186v2
- Date: Sun, 16 Aug 2020 00:07:39 GMT
- Title: University-1652: A Multi-view Multi-source Benchmark for Drone-based
Geo-localization
- Authors: Zhedong Zheng and Yunchao Wei and Yi Yang
- Abstract summary: We introduce a new multi-view benchmark for drone-based geo-localization, named University-1652.
University-1652 contains data from three platforms, i.e., synthetic drones, satellites and ground cameras of 1,652 university buildings around the world.
Experiments show that University-1652 helps the model to learn the viewpoint-invariant features and also has good generalization ability in the real-world scenario.
- Score: 87.74121935246937
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of cross-view geo-localization. The primary challenge
of this task is to learn the robust feature against large viewpoint changes.
Existing benchmarks can help, but are limited in the number of viewpoints.
Image pairs, containing two viewpoints, e.g., satellite and ground, are usually
provided, which may compromise the feature learning. Besides phone cameras and
satellites, in this paper, we argue that drones could serve as the third
platform to deal with the geo-localization problem. In contrast to the
traditional ground-view images, drone-view images meet fewer obstacles, e.g.,
trees, and could provide a comprehensive view when flying around the target
place. To verify the effectiveness of the drone platform, we introduce a new
multi-view multi-source benchmark for drone-based geo-localization, named
University-1652. University-1652 contains data from three platforms, i.e.,
synthetic drones, satellites and ground cameras of 1,652 university buildings
around the world. To our knowledge, University-1652 is the first drone-based
geo-localization dataset and enables two new tasks, i.e., drone-view target
localization and drone navigation. As the name implies, drone-view target
localization intends to predict the location of the target place via drone-view
images. On the other hand, given a satellite-view query image, drone navigation
is to drive the drone to the area of interest in the query. We use this dataset
to analyze a variety of off-the-shelf CNN features and propose a strong CNN
baseline on this challenging dataset. The experiments show that University-1652
helps the model to learn the viewpoint-invariant features and also has good
generalization ability in the real-world scenario.
Related papers
- Game4Loc: A UAV Geo-Localization Benchmark from Game Data [0.0]
We introduce a more practical UAV geo-localization task including partial matches of cross-view paired data.
Experiments demonstrate the effectiveness of our data and training method for UAV geo-localization.
arXiv Detail & Related papers (2024-09-25T13:33:28Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - TransVisDrone: Spatio-Temporal Transformer for Vision-based
Drone-to-Drone Detection in Aerial Videos [57.92385818430939]
Drone-to-drone detection using visual feed has crucial applications, such as detecting drone collisions, detecting drone attacks, or coordinating flight with other drones.
Existing methods are computationally costly, follow non-end-to-end optimization, and have complex multi-stage pipelines, making them less suitable for real-time deployment on edge devices.
We propose a simple yet effective framework, itTransVisDrone, that provides an end-to-end solution with higher computational efficiency.
arXiv Detail & Related papers (2022-10-16T03:05:13Z) - Geo-Localization via Ground-to-Satellite Cross-View Image Retrieval [25.93015219830576]
Given a ground-view image of a landmark, we aim to achieve cross-view geo-localization by searching out its corresponding satellite-view images.
We take advantage of drone-view information as a bridge between ground-view and satellite-view domains.
arXiv Detail & Related papers (2022-05-22T17:35:13Z) - SUES-200: A Multi-height Multi-scene Cross-view Image Benchmark Across
Drone and Satellite [0.0]
The purpose of cross-view image matching is to match images acquired from different platforms of the same target scene.
SUES-200 is the first dataset that considers the differences generated by aerial photography of drones at different flight heights.
arXiv Detail & Related papers (2022-04-22T13:49:52Z) - Dogfight: Detecting Drones from Drones Videos [58.158988162743825]
This paper attempts to address the problem of drones detection from other flying drones variations.
The erratic movement of the source and target drones, small size, arbitrary shape, large intensity, and occlusion make this problem quite challenging.
To handle this, instead of using region-proposal based methods, we propose to use a two-stage segmentation-based approach.
arXiv Detail & Related papers (2021-03-31T17:43:31Z) - Multi-Drone based Single Object Tracking with Agent Sharing Network [74.8198920355117]
Multi-Drone single Object Tracking dataset consists of 92 groups of video clips with 113,918 high resolution frames taken by two drones and 63 groups of video clips with 145,875 high resolution frames taken by three drones.
Agent sharing network (ASNet) is proposed by self-supervised template sharing and view-aware fusion of the target from multiple drones.
arXiv Detail & Related papers (2020-03-16T03:27:04Z) - Detection and Tracking Meet Drones Challenge [131.31749447313197]
This paper presents a review of object detection and tracking datasets and benchmarks, and discusses the challenges of collecting large-scale drone-based object detection and tracking datasets with manual annotations.
We describe our VisDrone dataset, which is captured over various urban/suburban areas of 14 different cities across China from North to South.
We provide a detailed analysis of the current state of the field of large-scale object detection and tracking on drones, and conclude the challenge as well as propose future directions.
arXiv Detail & Related papers (2020-01-16T00:11:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.