On the Advantages of Multiple Stereo Vision Camera Designs for
Autonomous Drone Navigation
- URL: http://arxiv.org/abs/2105.12691v1
- Date: Wed, 26 May 2021 17:10:20 GMT
- Title: On the Advantages of Multiple Stereo Vision Camera Designs for
Autonomous Drone Navigation
- Authors: Rui Pimentel de Figueiredo, Jakob Grimm Hansen, Jonas Le Fevre, Martim
Brand\~ao, Erdal Kayacan
- Abstract summary: We showcase the performance of a multi-camera UAV, when coupled with state-of-the-art planning and mapping algorithms.
We employ our approaches in an autonomous drone-based inspection task and evaluate them in an autonomous exploration and mapping scenario.
- Score: 7.299239909796724
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work we showcase the design and assessment of the performance of a
multi-camera UAV, when coupled with state-of-the-art planning and mapping
algorithms for autonomous navigation. The system leverages state-of-the-art
receding horizon exploration techniques for Next-Best-View (NBV) planning with
3D and semantic information, provided by a reconfigurable multi stereo camera
system. We employ our approaches in an autonomous drone-based inspection task
and evaluate them in an autonomous exploration and mapping scenario. We discuss
the advantages and limitations of using multi stereo camera flying systems, and
the trade-off between number of cameras and mapping performance.
Related papers
- Enhancing Autonomous Navigation by Imaging Hidden Objects using Single-Photon LiDAR [12.183773707869069]
We present a novel approach that leverages Non-Line-of-Sight (NLOS) sensing using single-photon LiDAR to improve visibility and enhance autonomous navigation.
Our method enables mobile robots to "see around corners" by utilizing multi-bounce light information.
arXiv Detail & Related papers (2024-10-04T16:03:13Z) - Design and Flight Demonstration of a Quadrotor for Urban Mapping and Target Tracking Research [0.04712282770819683]
This paper describes the hardware design and flight demonstration of a small quadrotor with imaging sensors for urban mapping, hazard avoidance, and target tracking research.
The vehicle is equipped with five cameras, including two pairs of fisheye stereo cameras that enable a nearly omnidirectional view and a two-axis gimbaled camera.
An onboard NVIDIA Jetson Orin Nano computer running the Robot Operating System software is used for data collection.
arXiv Detail & Related papers (2024-02-20T18:06:00Z) - Tri-Perspective View for Vision-Based 3D Semantic Occupancy Prediction [84.94140661523956]
We propose a tri-perspective view (TPV) representation which accompanies BEV with two additional perpendicular planes.
We model each point in the 3D space by summing its projected features on the three planes.
Experiments show that our model trained with sparse supervision effectively predicts the semantic occupancy for all voxels.
arXiv Detail & Related papers (2023-02-15T17:58:10Z) - A Simple Baseline for Multi-Camera 3D Object Detection [94.63944826540491]
3D object detection with surrounding cameras has been a promising direction for autonomous driving.
We present SimMOD, a Simple baseline for Multi-camera Object Detection.
We conduct extensive experiments on the 3D object detection benchmark of nuScenes to demonstrate the effectiveness of SimMOD.
arXiv Detail & Related papers (2022-08-22T03:38:01Z) - BEVerse: Unified Perception and Prediction in Birds-Eye-View for
Vision-Centric Autonomous Driving [92.05963633802979]
We present BEVerse, a unified framework for 3D perception and prediction based on multi-camera systems.
We show that the multi-task BEVerse outperforms single-task methods on 3D object detection, semantic map construction, and motion prediction.
arXiv Detail & Related papers (2022-05-19T17:55:35Z) - Scalable and Real-time Multi-Camera Vehicle Detection,
Re-Identification, and Tracking [58.95210121654722]
We propose a real-time city-scale multi-camera vehicle tracking system that handles real-world, low-resolution CCTV instead of idealized and curated video streams.
Our method is ranked among the top five performers on the public leaderboard.
arXiv Detail & Related papers (2022-04-15T12:47:01Z) - SurroundDepth: Entangling Surrounding Views for Self-Supervised
Multi-Camera Depth Estimation [101.55622133406446]
We propose a SurroundDepth method to incorporate the information from multiple surrounding views to predict depth maps across cameras.
Specifically, we employ a joint network to process all the surrounding views and propose a cross-view transformer to effectively fuse the information from multiple views.
In experiments, our method achieves the state-of-the-art performance on the challenging multi-camera depth estimation datasets.
arXiv Detail & Related papers (2022-04-07T17:58:47Z) - MFuseNet: Robust Depth Estimation with Learned Multiscopic Fusion [47.2251122861135]
We design a multiscopic vision system that utilizes a low-cost monocular RGB camera to acquire accurate depth estimation.
Unlike multi-view stereo with images captured at unconstrained camera poses, the proposed system controls the motion of a camera to capture a sequence of images.
arXiv Detail & Related papers (2021-08-05T08:31:01Z) - Understanding Bird's-Eye View Semantic HD-Maps Using an Onboard
Monocular Camera [110.83289076967895]
We study scene understanding in the form of online estimation of semantic bird's-eye-view HD-maps using the video input from a single onboard camera.
In our experiments, we demonstrate that the considered aspects are complementary to each other for HD-map understanding.
arXiv Detail & Related papers (2020-12-05T14:39:14Z) - Do You See What I See? Coordinating Multiple Aerial Cameras for Robot
Cinematography [9.870369982132678]
We develop a real-time multi-UAV coordination system that is capable of recording dynamic targets while maximizing shot diversity and avoiding collisions.
We show that our coordination scheme has low computational cost and takes only 1.17 ms on average to plan for a team of 3 UAVs over a 10 s time horizon.
arXiv Detail & Related papers (2020-11-10T22:43:25Z) - Deep Multimodality Learning for UAV Video Aesthetic Quality Assessment [22.277636020333198]
We present a method of deep multimodality learning for UAV video aesthetic quality assessment.
A novel specially designed motion stream network is proposed for this new multistream framework.
We present three application examples of UAV video grading, professional segment detection and aesthetic-based UAV path planning.
arXiv Detail & Related papers (2020-11-04T15:37:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.