Learning Model Predictive Controllers with Real-Time Attention for
Real-World Navigation
- URL: http://arxiv.org/abs/2209.10780v2
- Date: Sat, 24 Sep 2022 01:12:08 GMT
- Title: Learning Model Predictive Controllers with Real-Time Attention for
Real-World Navigation
- Authors: Xuesu Xiao, Tingnan Zhang, Krzysztof Choromanski, Edward Lee, Anthony
Francis, Jake Varley, Stephen Tu, Sumeet Singh, Peng Xu, Fei Xia, Sven Mikael
Persson, Dmitry Kalashnikov, Leila Takayama, Roy Frostig, Jie Tan, Carolina
Parada, Vikas Sindhwani
- Abstract summary: We present a new class of implicit control policies combining the benefits of imitation learning with the robust handling of system constraints.
Our approach, called Performer-MPC, uses a learned cost function parameterized by vision context embeddings provided by Performers.
Compared with a standard MPC policy, Performer-MPC achieves >40% better goal reached in cluttered environments and >65% better on social metrics when navigating around humans.
- Score: 34.86856430694435
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite decades of research, existing navigation systems still face
real-world challenges when deployed in the wild, e.g., in cluttered home
environments or in human-occupied public spaces. To address this, we present a
new class of implicit control policies combining the benefits of imitation
learning with the robust handling of system constraints from Model Predictive
Control (MPC). Our approach, called Performer-MPC, uses a learned cost function
parameterized by vision context embeddings provided by Performers -- a low-rank
implicit-attention Transformer. We jointly train the cost function and
construct the controller relying on it, effectively solving end-to-end the
corresponding bi-level optimization problem. We show that the resulting policy
improves standard MPC performance by leveraging a few expert demonstrations of
the desired navigation behavior in different challenging real-world scenarios.
Compared with a standard MPC policy, Performer-MPC achieves >40% better goal
reached in cluttered environments and >65% better on social metrics when
navigating around humans.
Related papers
- Parameter-Adaptive Approximate MPC: Tuning Neural-Network Controllers without Retraining [50.00291020618743]
This work introduces a novel, parameter-adaptive AMPC architecture capable of online tuning without recomputing large datasets and retraining.
We showcase the effectiveness of parameter-adaptive AMPC by controlling the swing-ups of two different real cartpole systems with a severely resource-constrained microcontroller (MCU)
Taken together, these contributions represent a marked step toward the practical application of AMPC in real-world systems.
arXiv Detail & Related papers (2024-04-08T20:02:19Z) - Reinforcement Learning with Model Predictive Control for Highway Ramp Metering [14.389086937116582]
This work explores the synergy between model-based and learning-based strategies to enhance traffic flow management.
The control problem is formulated as an RL task by crafting a suitable stage cost function.
An MPC-based RL approach, which leverages the MPC optimal problem as a function approximation for the RL algorithm, is proposed to learn to efficiently control an on-ramp.
arXiv Detail & Related papers (2023-11-15T09:50:54Z) - GP-guided MPPI for Efficient Navigation in Complex Unknown Cluttered
Environments [2.982218441172364]
This study presents the GP-MPPI, an online learning-based control strategy that integrates Model Predictive Path Intergal (MPPI) with a local perception model.
We validate the efficiency and robustness of our proposed control strategy through both simulated and real-world experiments of 2D autonomous navigation tasks.
arXiv Detail & Related papers (2023-07-08T17:33:20Z) - GAN-MPC: Training Model Predictive Controllers with Parameterized Cost
Functions using Demonstrations from Non-identical Experts [14.291720751625585]
We propose a generative adversarial network (GAN) to minimize the Jensen-Shannon divergence between the state-trajectory distributions of the demonstrator and the imitator.
We evaluate our approach on a variety of simulated robotics tasks of DeepMind Control suite.
arXiv Detail & Related papers (2023-05-30T15:15:30Z) - Efficient Deep Learning of Robust, Adaptive Policies using Tube
MPC-Guided Data Augmentation [42.66792060626531]
Existing robust and adaptive controllers can achieve impressive performance at the cost of heavy online onboard computations.
We extend an existing efficient Imitation Learning (IL) algorithm for robust policy learning from MPC with the ability to learn policies that adapt to challenging model/environment uncertainties.
arXiv Detail & Related papers (2023-03-28T02:22:47Z) - Policy Search for Model Predictive Control with Application to Agile
Drone Flight [56.24908013905407]
We propose a policy-search-for-model-predictive-control framework for MPC.
Specifically, we formulate the MPC as a parameterized controller, where the hard-to-optimize decision variables are represented as high-level policies.
Experiments show that our controller achieves robust and real-time control performance in both simulation and the real world.
arXiv Detail & Related papers (2021-12-07T17:39:24Z) - Imitation Learning from MPC for Quadrupedal Multi-Gait Control [63.617157490920505]
We present a learning algorithm for training a single policy that imitates multiple gaits of a walking robot.
We use and extend MPC-Net, which is an Imitation Learning approach guided by Model Predictive Control.
We validate our approach on hardware and show that a single learned policy can replace its teacher to control multiple gaits.
arXiv Detail & Related papers (2021-03-26T08:48:53Z) - Learning High-Level Policies for Model Predictive Control [54.00297896763184]
Model Predictive Control (MPC) provides robust solutions to robot control tasks.
We propose a self-supervised learning algorithm for learning a neural network high-level policy.
We show that our approach can handle situations that are difficult for standard MPC.
arXiv Detail & Related papers (2020-07-20T17:12:34Z) - Guided Constrained Policy Optimization for Dynamic Quadrupedal Robot
Locomotion [78.46388769788405]
We introduce guided constrained policy optimization (GCPO), an RL framework based upon our implementation of constrained policy optimization (CPPO)
We show that guided constrained RL offers faster convergence close to the desired optimum resulting in an optimal, yet physically feasible, robotic control behavior without the need for precise reward function tuning.
arXiv Detail & Related papers (2020-02-22T10:15:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.