Aerial-PASS: Panoramic Annular Scene Segmentation in Drone Videos
- URL: http://arxiv.org/abs/2105.07209v1
- Date: Sat, 15 May 2021 12:01:16 GMT
- Title: Aerial-PASS: Panoramic Annular Scene Segmentation in Drone Videos
- Authors: Lei Sun, Jia Wang, Kailun Yang, Kaikai Wu, Xiangdong Zhou, Kaiwei
Wang, Jian Bai
- Abstract summary: We design a UAV system with a Panoramic Annular Lens (PAL), which has the characteristics of small size, low weight, and a 360-degree annular FoV.
A lightweight panoramic annular semantic segmentation neural network model is designed to achieve high-accuracy and real-time scene parsing.
A comprehensive variety of experiments shows that the designed system performs satisfactorily in aerial panoramic scene parsing.
- Score: 15.244418294614857
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Aerial pixel-wise scene perception of the surrounding environment is an
important task for UAVs (Unmanned Aerial Vehicles). Previous research works
mainly adopt conventional pinhole cameras or fisheye cameras as the imaging
device. However, these imaging systems cannot achieve large Field of View
(FoV), small size, and lightweight at the same time. To this end, we design a
UAV system with a Panoramic Annular Lens (PAL), which has the characteristics
of small size, low weight, and a 360-degree annular FoV. A lightweight
panoramic annular semantic segmentation neural network model is designed to
achieve high-accuracy and real-time scene parsing. In addition, we present the
first drone-perspective panoramic scene segmentation dataset Aerial-PASS, with
annotated labels of track, field, and others. A comprehensive variety of
experiments shows that the designed system performs satisfactorily in aerial
panoramic scene parsing. In particular, our proposed model strikes an excellent
trade-off between segmentation performance and inference speed suitable,
validated on both public street-scene and our established aerial-scene
datasets.
Related papers
- Radiance Field Learners As UAV First-Person Viewers [36.59524833437512]
First-Person-View (FPV) holds immense potential for revolutionizing the trajectory of Unmanned Aerial Vehicles (UAVs)
Traditional Neural Radiance Field (NeNeRF) methods face challenges such as sampling single points per granularity.
We introduce FPV-NeRF, addressing these challenges through three key facets.
arXiv Detail & Related papers (2024-08-10T12:29:11Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - Towards Viewpoint Robustness in Bird's Eye View Segmentation [85.99907496019972]
We study how AV perception models are affected by changes in camera viewpoint.
Small changes to pitch, yaw, depth, or height of the camera at inference time lead to large drops in performance.
We introduce a technique for novel view synthesis and use it to transform collected data to the viewpoint of target rigs.
arXiv Detail & Related papers (2023-09-11T02:10:07Z) - Tri-Perspective View for Vision-Based 3D Semantic Occupancy Prediction [84.94140661523956]
We propose a tri-perspective view (TPV) representation which accompanies BEV with two additional perpendicular planes.
We model each point in the 3D space by summing its projected features on the three planes.
Experiments show that our model trained with sparse supervision effectively predicts the semantic occupancy for all voxels.
arXiv Detail & Related papers (2023-02-15T17:58:10Z) - Monocular BEV Perception of Road Scenes via Front-to-Top View Projection [57.19891435386843]
We present a novel framework that reconstructs a local map formed by road layout and vehicle occupancy in the bird's-eye view.
Our model runs at 25 FPS on a single GPU, which is efficient and applicable for real-time panorama HD map reconstruction.
arXiv Detail & Related papers (2022-11-15T13:52:41Z) - PL-EVIO: Robust Monocular Event-based Visual Inertial Odometry with
Point and Line Features [3.6355269783970394]
Event cameras are motion-activated sensors that capture pixel-level illumination changes instead of the intensity image with a fixed frame rate.
We propose a robust, high-accurate, and real-time optimization-based monocular event-based visual-inertial odometry (VIO) method.
arXiv Detail & Related papers (2022-09-25T06:14:12Z) - VPAIR -- Aerial Visual Place Recognition and Localization in Large-scale
Outdoor Environments [49.82314641876602]
We present a new dataset named VPAIR.
The dataset was recorded on board a light aircraft flying at an altitude of more than 300 meters above ground.
The dataset covers a more than one hundred kilometers long trajectory over various types of challenging landscapes.
arXiv Detail & Related papers (2022-05-23T18:50:08Z) - Agile wide-field imaging with selective high resolution [3.0080996413230667]
We report an agile wide-field imaging framework with selective high resolution that requires only two detectors.
Under this assumption, we use a short-focal camera to image wide field with a certain low resolution, and use a long-focal camera to acquire the HR images of ROI.
To automatically locate ROI in the wide field in real time, we propose an efficient deep-learning based multiscale registration method.
arXiv Detail & Related papers (2021-06-09T14:01:34Z) - Deep Multimodality Learning for UAV Video Aesthetic Quality Assessment [22.277636020333198]
We present a method of deep multimodality learning for UAV video aesthetic quality assessment.
A novel specially designed motion stream network is proposed for this new multistream framework.
We present three application examples of UAV video grading, professional segment detection and aesthetic-based UAV path planning.
arXiv Detail & Related papers (2020-11-04T15:37:49Z) - Perceiving Traffic from Aerial Images [86.994032967469]
We propose an object detection method called Butterfly Detector that is tailored to detect objects in aerial images.
We evaluate our Butterfly Detector on two publicly available UAV datasets (UAVDT and VisDrone 2019) and show that it outperforms previous state-of-the-art methods while remaining real-time.
arXiv Detail & Related papers (2020-09-16T11:37:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.