Brain over Brawn -- Using a Stereo Camera to Detect, Track and Intercept
a Faster UAV by Reconstructing Its Trajectory
- URL: http://arxiv.org/abs/2107.00962v1
- Date: Fri, 2 Jul 2021 10:49:22 GMT
- Title: Brain over Brawn -- Using a Stereo Camera to Detect, Track and Intercept
a Faster UAV by Reconstructing Its Trajectory
- Authors: Antonella Bari\v{s}i\'c, Frano Petric, Stjepan Bogdan
- Abstract summary: This paper demonstrates our approach to intercepting a faster intruder UAV, inspired by the MBZIRC 2020 Challenge 1.
By leveraging the shape of the intruder's trajectory we are able to calculate the interception point.
The system is able to track and intercept the target which is 30% faster than the interceptor in majority of simulation experiments.
- Score: 2.1638817206926855
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The work presented in this paper demonstrates our approach to intercepting a
faster intruder UAV, inspired by the MBZIRC2020 Challenge 1. By leveraging the
knowledge of the shape of the intruder's trajectory we are able to calculate
the interception point. Target tracking is based on image processing by a
YOLOv3 Tiny convolutional neural network, combined with depth calculation using
a gimbal-mounted ZED Mini stereo camera. We use RGB and depth data from ZED
Mini to extract the 3D position of the target, for which we devise a
histogram-of-depth based processing to reduce noise. Obtained 3D measurements
of target's position are used to calculate the position, the orientation and
the size of a figure-eight shaped trajectory, which we approximate using
lemniscate of Bernoulli. Once the approximation is deemed sufficiently precise,
measured by Hausdorff distance between measurements and the approximation, an
interception point is calculated to position the intercepting UAV right on the
path of the target. The proposed method, which has been significantly improved
based on the experience gathered during the MBZIRC competition, has been
validated in simulation and through field experiments. The results confirmed
that an efficient visual perception module which extracts information related
to the motion of the target UAV as a basis for the interception, has been
developed. The system is able to track and intercept the target which is 30%
faster than the interceptor in majority of simulation experiments. Tests in the
unstructured environment yielded 9 out of 12 successful results.
Related papers
- Revisiting Stereo Triangulation in UAV Distance Estimation [5.656973345209692]
We build and present a UAVDE dataset for UAV distance estimation, in which distance between two UAVs is obtained by UWB sensors.
We propose a novel position correction module, which can directly predict the offset between the observed positions and the actual ones.
We conduct extensive experiments on UAVDE, and our method can achieve a significant performance improvement over a strong baseline.
arXiv Detail & Related papers (2023-06-15T08:22:05Z) - RealNet: Combining Optimized Object Detection with Information Fusion
Depth Estimation Co-Design Method on IoT [2.9275056713717285]
We propose a co-design method combining the model-streamlined recognition algorithm, the depth estimation algorithm, and information fusion.
The method proposed in this paper is suitable for mobile platforms with high real-time request.
arXiv Detail & Related papers (2022-04-24T08:35:55Z) - iSDF: Real-Time Neural Signed Distance Fields for Robot Perception [64.80458128766254]
iSDF is a continuous learning system for real-time signed distance field reconstruction.
It produces more accurate reconstructions and better approximations of collision costs and gradients.
arXiv Detail & Related papers (2022-04-05T15:48:39Z) - RAANet: Range-Aware Attention Network for LiDAR-based 3D Object
Detection with Auxiliary Density Level Estimation [11.180128679075716]
Range-Aware Attention Network (RAANet) is developed for 3D object detection from LiDAR data for autonomous driving.
RAANet extracts more powerful BEV features and generates superior 3D object detections.
Experiments on nuScenes dataset demonstrate that our proposed approach outperforms the state-of-the-art methods for LiDAR-based 3D object detection.
arXiv Detail & Related papers (2021-11-18T04:20:13Z) - Trajectory Design for UAV-Based Internet-of-Things Data Collection: A
Deep Reinforcement Learning Approach [93.67588414950656]
In this paper, we investigate an unmanned aerial vehicle (UAV)-assisted Internet-of-Things (IoT) system in a 3D environment.
We present a TD3-based trajectory design for completion time minimization (TD3-TDCTM) algorithm.
Our simulation results show the superiority of the proposed TD3-TDCTM algorithm over three conventional non-learning based baseline methods.
arXiv Detail & Related papers (2021-07-23T03:33:29Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Deep Learning based Multi-Modal Sensing for Tracking and State
Extraction of Small Quadcopters [3.019035926889528]
This paper proposes a multi-sensor based approach to detect, track, and localize a quadcopter unmanned aerial vehicle (UAV)
Specifically, a pipeline is developed to process monocular RGB and thermal video (captured from a fixed platform) to detect and track the UAV in our FoV.
A 2D planar lidar is used to allow conversion of pixel data to actual distance measurements, and thereby enable localization of the UAV in global coordinates.
arXiv Detail & Related papers (2020-12-08T23:59:48Z) - Reinforced Axial Refinement Network for Monocular 3D Object Detection [160.34246529816085]
Monocular 3D object detection aims to extract the 3D position and properties of objects from a 2D input image.
Conventional approaches sample 3D bounding boxes from the space and infer the relationship between the target object and each of them, however, the probability of effective samples is relatively small in the 3D space.
We propose to start with an initial prediction and refine it gradually towards the ground truth, with only one 3d parameter changed in each step.
This requires designing a policy which gets a reward after several steps, and thus we adopt reinforcement learning to optimize it.
arXiv Detail & Related papers (2020-08-31T17:10:48Z) - Expandable YOLO: 3D Object Detection from RGB-D Images [64.14512458954344]
This paper aims at constructing a light-weight object detector that inputs a depth and a color image from a stereo camera.
By extending the network architecture of YOLOv3 to 3D in the middle, it is possible to output in the depth direction.
Intersection over Uninon (IoU) in 3D space is introduced to confirm the accuracy of region extraction results.
arXiv Detail & Related papers (2020-06-26T07:32:30Z) - End-to-end Learning for Inter-Vehicle Distance and Relative Velocity
Estimation in ADAS with a Monocular Camera [81.66569124029313]
We propose a camera-based inter-vehicle distance and relative velocity estimation method based on end-to-end training of a deep neural network.
The key novelty of our method is the integration of multiple visual clues provided by any two time-consecutive monocular frames.
We also propose a vehicle-centric sampling mechanism to alleviate the effect of perspective distortion in the motion field.
arXiv Detail & Related papers (2020-06-07T08:18:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.