SA-NET.v2: Real-time vehicle detection from oblique UAV images with use
of uncertainty estimation in deep meta-learning
- URL: http://arxiv.org/abs/2208.04190v1
- Date: Thu, 4 Aug 2022 09:08:47 GMT
- Title: SA-NET.v2: Real-time vehicle detection from oblique UAV images with use
of uncertainty estimation in deep meta-learning
- Authors: Mehdi Khoshboresh-Masouleh and Reza Shah-Hosseini
- Abstract summary: In this manuscript, we consider the problem of real-time vehicle detection for oblique UAV images based on a small training dataset and deep meta-learning.
The SA-Net.v2 is a developed method based on the SA-CNN for real-time vehicle detection by reformulating the squeeze-and-attention mechanism.
Experiments show that the SA-Net.v2 achieves promising performance in time series oblique UAV images.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, unmanned aerial vehicle (UAV) imaging is a suitable solution
for real-time monitoring different vehicles on the urban scale. Real-time
vehicle detection with the use of uncertainty estimation in deep meta-learning
for the portable platforms (e.g., UAV) potentially improves video understanding
in real-world applications with a small training dataset, while many vehicle
monitoring approaches appear to understand single-time detection with a big
training dataset. The purpose of real-time vehicle detection from oblique UAV
images is to locate the vehicle on the time series UAV images by using semantic
segmentation. Real-time vehicle detection is more difficult due to the variety
of depth and scale vehicles in oblique view UAV images. Motivated by these
facts, in this manuscript, we consider the problem of real-time vehicle
detection for oblique UAV images based on a small training dataset and deep
meta-learning. The proposed architecture, called SA-Net.v2, is a developed
method based on the SA-CNN for real-time vehicle detection by reformulating the
squeeze-and-attention mechanism. The SA-Net.v2 is composed of two components,
including the squeeze-and-attention function that extracts the high-level
feature based on a small training dataset, and the gated CNN. For the real-time
vehicle detection scenario, we test our model on the UAVid dataset. UAVid is a
time series oblique UAV images dataset consisting of 30 video sequences. We
examine the proposed method's applicability for stand real-time vehicle
detection in urban environments using time series UAV images. The experiments
show that the SA-Net.v2 achieves promising performance in time series oblique
UAV images.
Related papers
- EVD4UAV: An Altitude-Sensitive Benchmark to Evade Vehicle Detection in UAV [19.07281015014683]
Vehicle detection in Unmanned Aerial Vehicle (UAV) captured images has wide applications in aerial photography and remote sensing.
Recent studies show that adding an adversarial patch on objects can fool the well-trained deep neural networks based object detectors.
We propose a new dataset named EVD4UAV as an altitude-sensitive benchmark to evade vehicle detection in UAV.
arXiv Detail & Related papers (2024-03-08T16:19:39Z) - Real Time Human Detection by Unmanned Aerial Vehicles [0.0]
Two crucial data sources for public security are the thermal infrared (TIR) remote sensing photos and videos produced by unmanned aerial vehicles (UAVs)
due to the small scale of the target, complex scene information, low resolution relative to the viewable videos, and dearth of publicly available labeled datasets and training models, their object detection procedure is still difficult.
A UAV TIR object detection framework for pictures and videos is suggested in this study.
arXiv Detail & Related papers (2024-01-06T18:28:01Z) - Investigation of UAV Detection in Images with Complex Backgrounds and
Rainy Artifacts [20.20609511526255]
Vision-based object detection methods have been developed for UAV detection.
UAV detection in images with complex backgrounds and weather artifacts like rain has yet to be reasonably studied.
This work also focuses on benchmarking state-of-the-art object detection models.
arXiv Detail & Related papers (2023-05-25T19:54:33Z) - Deep Learning Computer Vision Algorithms for Real-time UAVs On-board
Camera Image Processing [77.34726150561087]
This paper describes how advanced deep learning based computer vision algorithms are applied to enable real-time on-board sensor processing for small UAVs.
All algorithms have been developed using state-of-the-art image processing methods based on deep neural networks.
arXiv Detail & Related papers (2022-11-02T11:10:42Z) - Analysis and Adaptation of YOLOv4 for Object Detection in Aerial Images [0.0]
Our work shows the adaptation of the popular YOLOv4 framework for predicting the objects and their locations in aerial images.
The trained model resulted in a mean average precision (mAP) of 45.64% with an inference speed reaching 8.7 FPS on the Tesla K80 GPU.
A comparative study with several contemporary aerial object detectors proved that YOLOv4 performed better, implying a more suitable detection algorithm to incorporate on aerial platforms.
arXiv Detail & Related papers (2022-03-18T23:51:09Z) - Real Time Monocular Vehicle Velocity Estimation using Synthetic Data [78.85123603488664]
We look at the problem of estimating the velocity of road vehicles from a camera mounted on a moving car.
We propose a two-step approach where first an off-the-shelf tracker is used to extract vehicle bounding boxes and then a small neural network is used to regress the vehicle velocity.
arXiv Detail & Related papers (2021-09-16T13:10:27Z) - Perceiving Traffic from Aerial Images [86.994032967469]
We propose an object detection method called Butterfly Detector that is tailored to detect objects in aerial images.
We evaluate our Butterfly Detector on two publicly available UAV datasets (UAVDT and VisDrone 2019) and show that it outperforms previous state-of-the-art methods while remaining real-time.
arXiv Detail & Related papers (2020-09-16T11:37:43Z) - V2VNet: Vehicle-to-Vehicle Communication for Joint Perception and
Prediction [74.42961817119283]
We use vehicle-to-vehicle (V2V) communication to improve the perception and motion forecasting performance of self-driving vehicles.
By intelligently aggregating the information received from multiple nearby vehicles, we can observe the same scene from different viewpoints.
arXiv Detail & Related papers (2020-08-17T17:58:26Z) - Vehicle Position Estimation with Aerial Imagery from Unmanned Aerial
Vehicles [4.555256739812733]
This work describes a process to estimate a precise vehicle position from aerial imagery.
The state-of-the-art deep neural network Mask-RCNN is applied for that purpose.
A mean accuracy of 20 cm can be achieved with flight altitudes up to 100 m, Full-HD resolution and a frame-by-frame detection.
arXiv Detail & Related papers (2020-04-17T12:29:40Z) - VehicleNet: Learning Robust Visual Representation for Vehicle
Re-identification [116.1587709521173]
We propose to build a large-scale vehicle dataset (called VehicleNet) by harnessing four public vehicle datasets.
We design a simple yet effective two-stage progressive approach to learning more robust visual representation from VehicleNet.
We achieve the state-of-art accuracy of 86.07% mAP on the private test set of AICity Challenge.
arXiv Detail & Related papers (2020-04-14T05:06:38Z) - Road Curb Detection and Localization with Monocular Forward-view Vehicle
Camera [74.45649274085447]
We propose a robust method for estimating road curb 3D parameters using a calibrated monocular camera equipped with a fisheye lens.
Our approach is able to estimate the vehicle to curb distance in real time with mean accuracy of more than 90%.
arXiv Detail & Related papers (2020-02-28T00:24:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.