MorphEyes: Variable Baseline Stereo For Quadrotor Navigation
- URL: http://arxiv.org/abs/2011.03077v1
- Date: Thu, 5 Nov 2020 20:04:35 GMT
- Title: MorphEyes: Variable Baseline Stereo For Quadrotor Navigation
- Authors: Nitin J. Sanket, Chahat Deep Singh, Varun Asthana, Cornelia
Ferm\"uller, Yiannis Aloimonos
- Abstract summary: We present a framework for quadrotor navigation based on a stereo camera system whose baseline can be adapted on-the-fly.
We show that our variable baseline system is more accurate and robust in all three scenarios.
- Score: 13.830987813403018
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Morphable design and depth-based visual control are two upcoming trends
leading to advancements in the field of quadrotor autonomy. Stereo-cameras have
struck the perfect balance of weight and accuracy of depth estimation but
suffer from the problem of depth range being limited and dictated by the
baseline chosen at design time. In this paper, we present a framework for
quadrotor navigation based on a stereo camera system whose baseline can be
adapted on-the-fly. We present a method to calibrate the system at a small
number of discrete baselines and interpolate the parameters for the entire
baseline range. We present an extensive theoretical analysis of calibration and
synchronization errors. We showcase three different applications of such a
system for quadrotor navigation: (a) flying through a forest, (b) flying
through an unknown shaped/location static/dynamic gap, and (c) accurate 3D pose
detection of an independently moving object. We show that our variable baseline
system is more accurate and robust in all three scenarios. To our knowledge,
this is the first work that applies the concept of morphable design to achieve
a variable baseline stereo vision system on a quadrotor.
Related papers
- GEOcc: Geometrically Enhanced 3D Occupancy Network with Implicit-Explicit Depth Fusion and Contextual Self-Supervision [49.839374549646884]
This paper presents GEOcc, a Geometric-Enhanced Occupancy network tailored for vision-only surround-view perception.
Our approach achieves State-Of-The-Art performance on the Occ3D-nuScenes dataset with the least image resolution needed and the most weightless image backbone.
arXiv Detail & Related papers (2024-05-17T07:31:20Z) - Vision-based control for landing an aerial vehicle on a marine vessel [0.0]
This work addresses the landing problem of an aerial vehicle, exemplified by a simple quadrotor, on a moving platform using image-based visual servo control.
The image features on the textured target plane are exploited to derive a vision-based control law.
The proposed control law guarantees convergence without estimating the unknown distance between the target and the moving platform.
arXiv Detail & Related papers (2024-04-17T12:53:57Z) - CHOSEN: Contrastive Hypothesis Selection for Multi-View Depth Refinement [17.4479165692548]
CHOSEN is a flexible, robust and effective multi-view depth refinement framework.
It can be employed in any existing multi-view stereo pipeline.
arXiv Detail & Related papers (2024-04-02T18:27:03Z) - OCTraN: 3D Occupancy Convolutional Transformer Network in Unstructured
Traffic Scenarios [0.0]
We propose OCTraN, a transformer architecture that uses iterative-attention to convert 2D image features into 3D occupancy features.
We also develop a self-supervised training pipeline to generalize the model to any scene by eliminating the need for LiDAR ground truth.
arXiv Detail & Related papers (2023-07-20T15:06:44Z) - Tightly-Coupled LiDAR-Visual SLAM Based on Geometric Features for Mobile
Agents [43.137917788594926]
We propose a tightly-coupled LiDAR-visual SLAM based on geometric features.
The entire line segment detected by the visual subsystem overcomes the limitation of the LiDAR subsystem.
Our system achieves more accurate and robust pose estimation compared to current state-of-the-art multi-modal methods.
arXiv Detail & Related papers (2023-07-15T10:06:43Z) - Geometric-aware Pretraining for Vision-centric 3D Object Detection [77.7979088689944]
We propose a novel geometric-aware pretraining framework called GAPretrain.
GAPretrain serves as a plug-and-play solution that can be flexibly applied to multiple state-of-the-art detectors.
We achieve 46.2 mAP and 55.5 NDS on the nuScenes val set using the BEVFormer method, with a gain of 2.7 and 2.1 points, respectively.
arXiv Detail & Related papers (2023-04-06T14:33:05Z) - Fully Convolutional One-Stage 3D Object Detection on LiDAR Range Images [96.66271207089096]
FCOS-LiDAR is a fully convolutional one-stage 3D object detector for LiDAR point clouds of autonomous driving scenes.
We show that an RV-based 3D detector with standard 2D convolutions alone can achieve comparable performance to state-of-the-art BEV-based detectors.
arXiv Detail & Related papers (2022-05-27T05:42:16Z) - Visual-tactile sensing for Real-time liquid Volume Estimation in
Grasping [58.50342759993186]
We propose a visuo-tactile model for realtime estimation of the liquid inside a deformable container.
We fuse two sensory modalities, i.e., the raw visual inputs from the RGB camera and the tactile cues from our specific tactile sensor.
The robotic system is well controlled and adjusted based on the estimation model in real time.
arXiv Detail & Related papers (2022-02-23T13:38:31Z) - SGM3D: Stereo Guided Monocular 3D Object Detection [62.11858392862551]
We propose a stereo-guided monocular 3D object detection network, termed SGM3D.
We exploit robust 3D features extracted from stereo images to enhance the features learned from the monocular image.
Our method can be integrated into many other monocular approaches to boost performance without introducing any extra computational cost.
arXiv Detail & Related papers (2021-12-03T13:57:14Z) - OmniSLAM: Omnidirectional Localization and Dense Mapping for
Wide-baseline Multi-camera Systems [88.41004332322788]
We present an omnidirectional localization and dense mapping system for a wide-baseline multiview stereo setup with ultra-wide field-of-view (FOV) fisheye cameras.
For more practical and accurate reconstruction, we first introduce improved and light-weighted deep neural networks for the omnidirectional depth estimation.
We integrate our omnidirectional depth estimates into the visual odometry (VO) and add a loop closing module for global consistency.
arXiv Detail & Related papers (2020-03-18T05:52:10Z) - Spatiotemporal Camera-LiDAR Calibration: A Targetless and Structureless
Approach [32.15405927679048]
We propose a targetless and structureless camera-DAR calibration method.
Our method combines a closed-form solution with a structureless bundle where the coarse-to-fine approach does not require an initial adjustment on the temporal parameters.
We demonstrate the accuracy and robustness of the proposed method through both simulation and real data experiments.
arXiv Detail & Related papers (2020-01-17T07:25:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.