Reinforcement learning of optimal active particle navigation
- URL: http://arxiv.org/abs/2202.00812v1
- Date: Tue, 1 Feb 2022 23:47:59 GMT
- Title: Reinforcement learning of optimal active particle navigation
- Authors: Mahdi Nasiri, Benno Liebchen
- Abstract summary: We develop a machine learning-based approach that allows us to determine the gradientally optimal path of a self-propelled agent.
Our method hinges on policy-based deep learning reinforcement techniques and, crucially, does not require any reward shaping or calculates.
The presented method provides a powerful alternative to current analytical methods and opens a route towards a universal path planner for future intelligent particles.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The development of self-propelled particles at the micro- and the nanoscale
has sparked a huge potential for future applications in active matter physics,
microsurgery, and targeted drug delivery. However, while the latter
applications provoke the quest on how to optimally navigate towards a target,
such as e.g. a cancer cell, there is still no simple way known to determine the
optimal route in sufficiently complex environments. Here we develop a machine
learning-based approach that allows us, for the first time, to determine the
asymptotically optimal path of a self-propelled agent which can freely steer in
complex environments. Our method hinges on policy gradient-based deep
reinforcement learning techniques and, crucially, does not require any reward
shaping or heuristics. The presented method provides a powerful alternative to
current analytical methods to calculate optimal trajectories and opens a route
towards a universal path planner for future intelligent active particles.
Related papers
- Research on Autonomous Robots Navigation based on Reinforcement Learning [13.559881645869632]
We use the Deep Q Network (DQN) and Proximal Policy Optimization (PPO) models to optimize the path planning and decision-making process.
We have verified the effectiveness and robustness of these models in various complex scenarios.
arXiv Detail & Related papers (2024-07-02T00:44:06Z) - Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning [53.3760591018817]
We propose a new benchmarking environment for aquatic navigation using recent advances in the integration between game engines and Deep Reinforcement Learning.
Specifically, we focus on PPO, one of the most widely accepted algorithms, and we propose advanced training techniques.
Our empirical evaluation shows that a well-designed combination of these ingredients can achieve promising results.
arXiv Detail & Related papers (2024-05-30T23:20:23Z) - Accelerating optimization over the space of probability measures [17.32262527237843]
We introduce a Hamiltonian-flow approach analogous to momentum-based approaches in Euclidean space.
We demonstrate that, in the continuous-time setting, algorithms based on this approach can achieve convergence rates of arbitrarily high order.
arXiv Detail & Related papers (2023-10-06T04:32:15Z) - Reparameterized Policy Learning for Multimodal Trajectory Optimization [61.13228961771765]
We investigate the challenge of parametrizing policies for reinforcement learning in high-dimensional continuous action spaces.
We propose a principled framework that models the continuous RL policy as a generative model of optimal trajectories.
We present a practical model-based RL method, which leverages the multimodal policy parameterization and learned world model.
arXiv Detail & Related papers (2023-07-20T09:05:46Z) - Optimal active particle navigation meets machine learning [0.0]
"Smart" active agents, like colloidal insects, microorganisms, or future robots, need to steer to optimally reach or discover a target, such as an odor source, food, or a cancer cell in a complex environment.
Here, we provide an overview of recent developments, regarding such optimal navigation problems, from the micro- to the macroscale.
arXiv Detail & Related papers (2023-03-09T19:48:03Z) - Reinforcement Learning for Molecular Dynamics Optimization: A Stochastic Pontryagin Maximum Principle Approach [3.0077933778535706]
We present a novel reinforcement learning framework designed to optimize molecular dynamics.
Our framework focuses on the entire trajectory rather than just the final molecular configuration.
Our method makes it suitable for applications in areas such as drug discovery and molecular design.
arXiv Detail & Related papers (2022-12-06T20:44:24Z) - Exploration via Planning for Information about the Optimal Trajectory [67.33886176127578]
We develop a method that allows us to plan for exploration while taking the task and the current knowledge into account.
We demonstrate that our method learns strong policies with 2x fewer samples than strong exploration baselines.
arXiv Detail & Related papers (2022-10-06T20:28:55Z) - Online reinforcement learning with sparse rewards through an active
inference capsule [62.997667081978825]
This paper introduces an active inference agent which minimizes the novel free energy of the expected future.
Our model is capable of solving sparse-reward problems with a very high sample efficiency.
We also introduce a novel method for approximating the prior model from the reward function, which simplifies the expression of complex objectives.
arXiv Detail & Related papers (2021-06-04T10:03:36Z) - Autonomous Drone Racing with Deep Reinforcement Learning [39.757652701917166]
In many robotic tasks, such as drone racing, the goal is to travel through a set of waypoints as fast as possible.
A key challenge is planning the minimum-time trajectory, which is typically solved by assuming perfect knowledge of the waypoints to pass in advance.
In this work, a new approach to minimum-time trajectory generation for quadrotors is presented.
arXiv Detail & Related papers (2021-03-15T18:05:49Z) - Path Planning Followed by Kinodynamic Smoothing for Multirotor Aerial
Vehicles (MAVs) [61.94975011711275]
We propose a geometrically based motion planning technique textquotedblleft RRT*textquotedblright; for this purpose.
In the proposed technique, we modified original RRT* introducing an adaptive search space and a steering function.
We have tested the proposed technique in various simulated environments.
arXiv Detail & Related papers (2020-08-29T09:55:49Z) - Localized active learning of Gaussian process state space models [63.97366815968177]
A globally accurate model is not required to achieve good performance in many common control applications.
We propose an active learning strategy for Gaussian process state space models that aims to obtain an accurate model on a bounded subset of the state-action space.
By employing model predictive control, the proposed technique integrates information collected during exploration and adaptively improves its exploration strategy.
arXiv Detail & Related papers (2020-05-04T05:35:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.