On Learning Vehicle Detection in Satellite Video
- URL: http://arxiv.org/abs/2001.10900v1
- Date: Wed, 29 Jan 2020 15:35:16 GMT
- Title: On Learning Vehicle Detection in Satellite Video
- Authors: Roman Pflugfelder, Axel Weissenfeld, Julian Wagner
- Abstract summary: Vehicle detection in aerial and satellite images is still challenging due to their tiny appearance in pixels compared to the overall size of remote sensing imagery.
This work proposes to apply recent work on deep learning for wide-area motion imagery (WAMI) on satellite video.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vehicle detection in aerial and satellite images is still challenging due to
their tiny appearance in pixels compared to the overall size of remote sensing
imagery. Classical methods of object detection very often fail in this scenario
due to violation of implicit assumptions made such as rich texture, small to
moderate ratios between image size and object size. Satellite video is a very
new modality which introduces temporal consistency as inductive bias.
Approaches for vehicle detection in satellite video use either background
subtraction, frame differencing or subspace methods showing moderate
performance (0.26 - 0.82 $F_1$ score). This work proposes to apply recent work
on deep learning for wide-area motion imagery (WAMI) on satellite video. We
show in a first approach comparable results (0.84 $F_1$) on Planet's SkySat-1
LasVegas video with room for further improvement.
Related papers
- Vehicle Perception from Satellite [54.07157185000604]
The dataset is constructed based on 12 satellite videos and 14 synthetic videos recorded from GTA-V.
It supports several tasks, including tiny object detection, counting and density estimation.
128,801 vehicles are annotated totally, and the number of vehicles in each image varies from 0 to 101.
arXiv Detail & Related papers (2024-02-01T15:59:16Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - DiffusionSat: A Generative Foundation Model for Satellite Imagery [63.2807119794691]
We present DiffusionSat, to date the largest generative foundation model trained on a collection of publicly available large, high-resolution remote sensing datasets.
Our method produces realistic samples and can be used to solve multiple generative tasks including temporal generation, superresolution given multi-spectral inputs and in-painting.
arXiv Detail & Related papers (2023-12-06T16:53:17Z) - CVLNet: Cross-View Semantic Correspondence Learning for Video-based
Camera Localization [89.69214577915959]
This paper tackles the problem of Cross-view Video-based camera localization.
We propose estimating the query camera's relative displacement to a satellite image before similarity matching.
Experiments have demonstrated the effectiveness of video-based localization over single image-based localization.
arXiv Detail & Related papers (2022-08-07T07:35:17Z) - Satellite Image Based Cross-view Localization for Autonomous Vehicle [59.72040418584396]
This paper shows that by using an off-the-shelf high-definition satellite image as a ready-to-use map, we are able to achieve cross-view vehicle localization up to a satisfactory accuracy.
Our method is validated on KITTI and Ford Multi-AV Seasonal datasets as ground view and Google Maps as the satellite view.
arXiv Detail & Related papers (2022-07-27T13:16:39Z) - Autonomous Satellite Detection and Tracking using Optical Flow [0.0]
An autonomous method of satellite detection and tracking in images is implemented using optical flow.
Optical flow is used to estimate the image velocities of detected objects in a series of space images.
The detection algorithm is exercised using both simulated star images and ground-based imagery of satellites.
arXiv Detail & Related papers (2022-04-14T15:23:27Z) - Deep Vehicle Detection in Satellite Video [0.0]
Vehicle detection is perhaps impossible in single satellite images due to the tininess of vehicles (4 pixel) and their similarity to the background EO.
A new model of a compact $3 times 3$ neural network is proposed which neglects pooling layers and uses leaky ReLUs.
Empirical results on two new annotated satellite videos reconfirm the applicability of this approach for vehicle detection.
arXiv Detail & Related papers (2022-04-14T08:54:44Z) - Beyond Cross-view Image Retrieval: Highly Accurate Vehicle Localization
Using Satellite Image [91.29546868637911]
This paper addresses the problem of vehicle-mounted camera localization by matching a ground-level image with an overhead-view satellite map.
The key idea is to formulate the task as pose estimation and solve it by neural-net based optimization.
Experiments on standard autonomous vehicle localization datasets have confirmed the superiority of the proposed method.
arXiv Detail & Related papers (2022-04-10T19:16:58Z) - Rapid Detection of Aircrafts in Satellite Imagery based on Deep Neural
Networks [2.7716102039510564]
This paper focuses on aircraft detection in satellite imagery using deep learning techniques.
In this paper, we used YOLO deep learning framework for aircraft detection.
The improved model shows good accuracy and performance on different unknown images having small, rotating, and dense objects to meet the requirements in real-time.
arXiv Detail & Related papers (2021-04-21T18:13:16Z) - Vehicle Position Estimation with Aerial Imagery from Unmanned Aerial
Vehicles [4.555256739812733]
This work describes a process to estimate a precise vehicle position from aerial imagery.
The state-of-the-art deep neural network Mask-RCNN is applied for that purpose.
A mean accuracy of 20 cm can be achieved with flight altitudes up to 100 m, Full-HD resolution and a frame-by-frame detection.
arXiv Detail & Related papers (2020-04-17T12:29:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.