Online No-regret Model-Based Meta RL for Personalized Navigation
- URL: http://arxiv.org/abs/2204.01925v1
- Date: Tue, 5 Apr 2022 01:28:06 GMT
- Title: Online No-regret Model-Based Meta RL for Personalized Navigation
- Authors: Yuda Song, Ye Yuan, Wen Sun, Kris Kitani
- Abstract summary: We propose an online no-regret model-based RL method that quickly conforms to the dynamics of the current user.
Our theoretical analysis shows that our method is a no-regret algorithm and we provide the convergence rate in the agnostic setting.
Our empirical analysis with 60+ hours of real-world user data shows that our method can reduce the number of collisions by more than 60%.
- Score: 37.82017324353145
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The interaction between a vehicle navigation system and the driver of the
vehicle can be formulated as a model-based reinforcement learning problem,
where the navigation systems (agent) must quickly adapt to the characteristics
of the driver (environmental dynamics) to provide the best sequence of
turn-by-turn driving instructions. Most modern day navigation systems (e.g,
Google maps, Waze, Garmin) are not designed to personalize their low-level
interactions for individual users across a wide range of driving styles (e.g.,
vehicle type, reaction time, level of expertise). Towards the development of
personalized navigation systems that adapt to a variety of driving styles, we
propose an online no-regret model-based RL method that quickly conforms to the
dynamics of the current user. As the user interacts with it, the navigation
system quickly builds a user-specific model, from which navigation commands are
optimized using model predictive control. By personalizing the policy in this
way, our method is able to give well-timed driving instructions that match the
user's dynamics. Our theoretical analysis shows that our method is a no-regret
algorithm and we provide the convergence rate in the agnostic setting. Our
empirical analysis with 60+ hours of real-world user data using a driving
simulator shows that our method can reduce the number of collisions by more
than 60%.
Related papers
- MetaFollower: Adaptable Personalized Autonomous Car Following [63.90050686330677]
We propose an adaptable personalized car-following framework - MetaFollower.
We first utilize Model-Agnostic Meta-Learning (MAML) to extract common driving knowledge from various CF events.
We additionally combine Long Short-Term Memory (LSTM) and Intelligent Driver Model (IDM) to reflect temporal heterogeneity with high interpretability.
arXiv Detail & Related papers (2024-06-23T15:30:40Z) - EditFollower: Tunable Car Following Models for Customizable Adaptive Cruise Control Systems [28.263763430300504]
We propose a data-driven car-following model that allows for adjusting driving discourtesy levels.
Our model provides valuable insights for the development of ACC systems that take into account drivers' social preferences.
arXiv Detail & Related papers (2024-06-23T15:04:07Z) - Avoidance Navigation Based on Offline Pre-Training Reinforcement
Learning [0.0]
This paper presents a Pre-Training Deep Reinforcement Learning(DRL) for avoidance navigation without map for mobile robots.
The efficient offline training strategy is proposed to speed up the inefficient random explorations in early stage.
It was demonstrated that our DRL model have universal general capacity in different environment.
arXiv Detail & Related papers (2023-08-03T06:19:46Z) - FastRLAP: A System for Learning High-Speed Driving via Deep RL and
Autonomous Practicing [71.76084256567599]
We present a system that enables an autonomous small-scale RC car to drive aggressively from visual observations using reinforcement learning (RL)
Our system, FastRLAP (faster lap), trains autonomously in the real world, without human interventions, and without requiring any simulation or expert demonstrations.
The resulting policies exhibit emergent aggressive driving skills, such as timing braking and acceleration around turns and avoiding areas which impede the robot's motion, approaching the performance of a human driver using a similar first-person interface over the course of training.
arXiv Detail & Related papers (2023-04-19T17:33:47Z) - Offline Reinforcement Learning for Visual Navigation [66.88830049694457]
ReViND is the first offline RL system for robotic navigation that can leverage previously collected data to optimize user-specified reward functions in the real-world.
We show that ReViND can navigate to distant goals using only offline training from this dataset, and exhibit behaviors that qualitatively differ based on the user-specified reward function.
arXiv Detail & Related papers (2022-12-16T02:23:50Z) - Tackling Real-World Autonomous Driving using Deep Reinforcement Learning [63.3756530844707]
In this work, we propose a model-free Deep Reinforcement Learning Planner training a neural network that predicts acceleration and steering angle.
In order to deploy the system on board the real self-driving car, we also develop a module represented by a tiny neural network.
arXiv Detail & Related papers (2022-07-05T16:33:20Z) - Learning to drive from a world on rails [78.28647825246472]
We learn an interactive vision-based driving policy from pre-recorded driving logs via a model-based approach.
A forward model of the world supervises a driving policy that predicts the outcome of any potential driving trajectory.
Our method ranks first on the CARLA leaderboard, attaining a 25% higher driving score while using 40 times less data.
arXiv Detail & Related papers (2021-05-03T05:55:30Z) - Autonomous Navigation through intersections with Graph
ConvolutionalNetworks and Conditional Imitation Learning for Self-driving
Cars [10.080958939027363]
In autonomous driving, navigation through unsignaled intersections is a challenging task.
We propose a novel branched network G-CIL for the navigation policy learning.
Our end-to-end trainable neural network outperforms the baselines with higher success rate and shorter navigation time.
arXiv Detail & Related papers (2021-02-01T07:33:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.