Autonomous Drone Racing with Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2103.08624v1
- Date: Mon, 15 Mar 2021 18:05:49 GMT
- Title: Autonomous Drone Racing with Deep Reinforcement Learning
- Authors: Yunlong Song, Mats Steinweg, Elia Kaufmann, and Davide Scaramuzza
- Abstract summary: In many robotic tasks, such as drone racing, the goal is to travel through a set of waypoints as fast as possible.
A key challenge is planning the minimum-time trajectory, which is typically solved by assuming perfect knowledge of the waypoints to pass in advance.
In this work, a new approach to minimum-time trajectory generation for quadrotors is presented.
- Score: 39.757652701917166
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In many robotic tasks, such as drone racing, the goal is to travel through a
set of waypoints as fast as possible. A key challenge for this task is planning
the minimum-time trajectory, which is typically solved by assuming perfect
knowledge of the waypoints to pass in advance. The resulting solutions are
either highly specialized for a single-track layout, or suboptimal due to
simplifying assumptions about the platform dynamics. In this work, a new
approach to minimum-time trajectory generation for quadrotors is presented.
Leveraging deep reinforcement learning and relative gate observations, this
approach can adaptively compute near-time-optimal trajectories for random track
layouts. Our method exhibits a significant computational advantage over
approaches based on trajectory optimization for non-trivial track
configurations. The proposed approach is evaluated on a set of race tracks in
simulation and the real world, achieving speeds of up to 17 m/s with a physical
quadrotor.
Related papers
- Revisiting Space Mission Planning: A Reinforcement Learning-Guided Approach for Multi-Debris Rendezvous [15.699822139827916]
The aim is to optimize the sequence in which all the given debris should be visited to get the least total time for rendezvous for the entire mission.
A neural network (NN) policy is developed, trained on simulated space missions with varying debris fields.
The reinforcement learning approach demonstrates a significant improvement in planning efficiency.
arXiv Detail & Related papers (2024-09-25T12:50:01Z) - Hybrid Imitation-Learning Motion Planner for Urban Driving [0.0]
We propose a novel hybrid motion planner that integrates both learning-based and optimization-based techniques.
Our model effectively balances safety and human-likeness, mitigating the trade-off inherent in these objectives.
We validate our approach through simulation experiments and further demonstrate its efficacy by deploying it in real-world self-driving vehicles.
arXiv Detail & Related papers (2024-09-04T16:54:31Z) - Evaluation of Local Planner-Based Stanley Control in Autonomous RC Car Racing Series [0.0]
This paper proposes a control technique for autonomous RC car racing.
It operates only local path planning on the actual LiDAR point cloud.
It is tested on a 1/10-sized RC car, and the tuning procedure from a base solution to the optimal setting in a real F1Tenth race is presented.
arXiv Detail & Related papers (2024-08-27T15:50:31Z) - Exploring Dynamic Transformer for Efficient Object Tracking [58.120191254379854]
We propose DyTrack, a dynamic transformer framework for efficient tracking.
DyTrack automatically learns to configure proper reasoning routes for various inputs, gaining better utilization of the available computational budget.
Experiments on multiple benchmarks demonstrate that DyTrack achieves promising speed-precision trade-offs with only a single model.
arXiv Detail & Related papers (2024-03-26T12:31:58Z) - Learning-Initialized Trajectory Planning in Unknown Environments [4.2960463890487555]
Planning for autonomous flight in unknown environments requires precise planning for both the spatial and temporal trajectories.
We introduce a novel approach that guides optimization using a Neural-d Trajectory Planner.
We propose a framework that supports robust online replanning with tolerance to planning latency.
arXiv Detail & Related papers (2023-09-19T15:07:26Z) - Efficient Ground Vehicle Path Following in Game AI [77.34726150561087]
This paper presents an efficient path following solution for ground vehicles tailored to game AI.
The proposed path follower is evaluated through a variety of test scenarios in a first-person shooter game.
We achieved a 70% decrease in the total number of stuck events compared to an existing path following solution.
arXiv Detail & Related papers (2023-07-07T04:20:07Z) - DDPEN: Trajectory Optimisation With Sub Goal Generation Model [70.36888514074022]
In this paper, we produce a novel Differential Dynamic Programming with Escape Network (DDPEN)
We propose to utilize a deep model that takes as an input map of the environment in the form of a costmap together with the desired position.
The model produces possible future directions that will lead to the goal, avoiding local minima which is possible to run in real time conditions.
arXiv Detail & Related papers (2023-01-18T11:02:06Z) - Time-Optimal Planning for Quadrotor Waypoint Flight [50.016821506107455]
Planning time-optimal trajectories at the actuation limit of a quadrotor is an open problem.
We propose a solution while exploiting the full quadrotor's actuator potential.
We validate our method in real-world flights in one of the world's largest motion-capture systems.
arXiv Detail & Related papers (2021-08-10T09:26:43Z) - Learning to Optimize Non-Rigid Tracking [54.94145312763044]
We employ learnable optimizations to improve robustness and speed up solver convergence.
First, we upgrade the tracking objective by integrating an alignment data term on deep features which are learned end-to-end through CNN.
Second, we bridge the gap between the preconditioning technique and learning method by introducing a ConditionNet which is trained to generate a preconditioner.
arXiv Detail & Related papers (2020-03-27T04:40:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.