DSR: Towards Drone Image Super-Resolution
- URL: http://arxiv.org/abs/2208.12327v1
- Date: Thu, 25 Aug 2022 19:58:54 GMT
- Title: DSR: Towards Drone Image Super-Resolution
- Authors: Xiaoyu Lin, Baran Ozaydin, Vidit Vidit, Majed El Helou and Sabine
S\"usstrunk
- Abstract summary: We propose a novel drone image dataset, with scenes captured at low and high resolutions, and across a span of altitudes.
Our results show that off-the-shelf state-of-the-art networks witness a significant drop in performance on this different domain.
We additionally show that simple fine-tuning, and incorporating altitude awareness into the network's architecture, both improve the reconstruction performance.
- Score: 10.679618027862846
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite achieving remarkable progress in recent years, single-image
super-resolution methods are developed with several limitations. Specifically,
they are trained on fixed content domains with certain degradations (whether
synthetic or real). The priors they learn are prone to overfitting the training
configuration. Therefore, the generalization to novel domains such as drone top
view data, and across altitudes, is currently unknown. Nonetheless, pairing
drones with proper image super-resolution is of great value. It would enable
drones to fly higher covering larger fields of view, while maintaining a high
image quality.
To answer these questions and pave the way towards drone image
super-resolution, we explore this application with particular focus on the
single-image case. We propose a novel drone image dataset, with scenes captured
at low and high resolutions, and across a span of altitudes. Our results show
that off-the-shelf state-of-the-art networks witness a significant drop in
performance on this different domain. We additionally show that simple
fine-tuning, and incorporating altitude awareness into the network's
architecture, both improve the reconstruction performance.
Related papers
- Scalable Indoor Novel-View Synthesis using Drone-Captured 360 Imagery with 3D Gaussian Splatting [3.020267522817212]
This paper proposes an efficient and scalable pipeline for indoor novel-view synthesis from drone-captured 360 videos.
360 cameras capture a wide set of viewpoints, allowing for comprehensive scene capture under a simple straightforward drone trajectory.
Our experiments demonstrate marked improvement in both reconstruction quality, i.e. PSNR and SSIM, and time compared to prior approaches.
arXiv Detail & Related papers (2024-10-15T05:08:47Z) - AG-NeRF: Attention-guided Neural Radiance Fields for Multi-height Large-scale Outdoor Scene Rendering [9.365775353436177]
Existing neural radiance fields (NeRF)-based novel view synthesis methods for large-scale outdoor scenes are mainly built on a single altitude.
We propose an end-to-end framework, termed AG-NeRF, and seek to reduce the training cost of building good reconstructions by synthesizing free-viewpoint images based on varying altitudes of scenes.
arXiv Detail & Related papers (2024-04-18T04:54:28Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - VDD: Varied Drone Dataset for Semantic Segmentation [9.581655974280217]
We release a large-scale, densely labeled collection of 400 high-resolution images spanning 7 classes.
This dataset features various scenes in urban, industrial, rural, and natural areas, captured from different camera angles and under diverse lighting conditions.
We train seven state-of-the-art models on drone datasets as baselines.
arXiv Detail & Related papers (2023-05-23T02:16:14Z) - Progressively Optimized Local Radiance Fields for Robust View Synthesis [76.55036080270347]
We present an algorithm for reconstructing the radiance field of a large-scale scene from a single casually captured video.
For handling unknown poses, we jointly estimate the camera poses with radiance field in a progressive manner.
For handling large unbounded scenes, we dynamically allocate new local radiance fields trained with frames within a temporal window.
arXiv Detail & Related papers (2023-03-24T04:03:55Z) - TransVisDrone: Spatio-Temporal Transformer for Vision-based
Drone-to-Drone Detection in Aerial Videos [57.92385818430939]
Drone-to-drone detection using visual feed has crucial applications, such as detecting drone collisions, detecting drone attacks, or coordinating flight with other drones.
Existing methods are computationally costly, follow non-end-to-end optimization, and have complex multi-stage pipelines, making them less suitable for real-time deployment on edge devices.
We propose a simple yet effective framework, itTransVisDrone, that provides an end-to-end solution with higher computational efficiency.
arXiv Detail & Related papers (2022-10-16T03:05:13Z) - Towards Robust Drone Vision in the Wild [1.14219428942199]
In this thesis, we propose the first image super-resolution dataset for drone vision.
We collect data at different altitudes and then propose pre-processing steps to align image pairs.
Two methods are proposed to build a robust image super-resolution network at different altitudes.
arXiv Detail & Related papers (2022-08-21T18:19:19Z) - SUES-200: A Multi-height Multi-scene Cross-view Image Benchmark Across
Drone and Satellite [0.0]
The purpose of cross-view image matching is to match images acquired from different platforms of the same target scene.
SUES-200 is the first dataset that considers the differences generated by aerial photography of drones at different flight heights.
arXiv Detail & Related papers (2022-04-22T13:49:52Z) - Aliasing is your Ally: End-to-End Super-Resolution from Raw Image Bursts [70.80220990106467]
This presentation addresses the problem of reconstructing a high-resolution image from multiple lower-resolution snapshots captured from slightly different viewpoints in space and time.
Key challenges for solving this problem include (i) aligning the input pictures with sub-pixel accuracy, (ii) handling raw (noisy) images for maximal faithfulness to native camera data, and (iii) designing/learning an image prior (regularizer) well suited to the task.
arXiv Detail & Related papers (2021-04-13T13:39:43Z) - Dogfight: Detecting Drones from Drones Videos [58.158988162743825]
This paper attempts to address the problem of drones detection from other flying drones variations.
The erratic movement of the source and target drones, small size, arbitrary shape, large intensity, and occlusion make this problem quite challenging.
To handle this, instead of using region-proposal based methods, we propose to use a two-stage segmentation-based approach.
arXiv Detail & Related papers (2021-03-31T17:43:31Z) - University-1652: A Multi-view Multi-source Benchmark for Drone-based
Geo-localization [87.74121935246937]
We introduce a new multi-view benchmark for drone-based geo-localization, named University-1652.
University-1652 contains data from three platforms, i.e., synthetic drones, satellites and ground cameras of 1,652 university buildings around the world.
Experiments show that University-1652 helps the model to learn the viewpoint-invariant features and also has good generalization ability in the real-world scenario.
arXiv Detail & Related papers (2020-02-27T15:24:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.