Optimization of the Model Predictive Control Update Interval Using
Reinforcement Learning
- URL: http://arxiv.org/abs/2011.13365v1
- Date: Thu, 26 Nov 2020 16:01:52 GMT
- Title: Optimization of the Model Predictive Control Update Interval Using
Reinforcement Learning
- Authors: Eivind B{\o}hn, Sebastien Gros, Signe Moe, Tor Arne Johansen
- Abstract summary: In control applications there is often a compromise that needs to be made with regards to the complexity and performance of the controller.
We propose a controller architecture in which the computational cost is explicitly optimized along with the control objective.
- Score: 0.7952582509792969
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In control applications there is often a compromise that needs to be made
with regards to the complexity and performance of the controller and the
computational resources that are available. For instance, the typical hardware
platform in embedded control applications is a microcontroller with limited
memory and processing power, and for battery powered applications the control
system can account for a significant portion of the energy consumption. We
propose a controller architecture in which the computational cost is explicitly
optimized along with the control objective. This is achieved by a three-part
architecture where a high-level, computationally expensive controller generates
plans, which a computationally simpler controller executes by compensating for
prediction errors, while a recomputation policy decides when the plan should be
recomputed. In this paper, we employ model predictive control (MPC) as the
high-level plan-generating controller, a linear state feedback controller as
the simpler compensating controller, and reinforcement learning (RL) to learn
the recomputation policy. Simulation results for two examples showcase the
architecture's ability to improve upon the MPC approach and find reasonable
compromises weighing the performance on the control objective and the
computational resources expended.
Related papers
- Goal-Conditioned Terminal Value Estimation for Real-time and Multi-task Model Predictive Control [1.2687745030755995]
We develop an MPC framework with goal-conditioned terminal value learning to achieve multitask policy optimization.
We evaluate the proposed method on a bipedal inverted pendulum robot model and confirm that combining goal-conditioned terminal value learning with an upper-level trajectory planner enables real-time control.
arXiv Detail & Related papers (2024-10-07T11:19:23Z) - Deployable Reinforcement Learning with Variable Control Rate [14.838483990647697]
We propose a variant of Reinforcement Learning (RL) with variable control rate.
In this approach, the policy decides the action the agent should take as well as the duration of the time step associated with that action.
We show the efficacy of SEAC through a proof-of-concept simulation driving an agent with Newtonian kinematics.
arXiv Detail & Related papers (2024-01-17T15:40:11Z) - MPC-Inspired Reinforcement Learning for Verifiable Model-Free Control [5.9867297878688195]
We introduce a new class of parameterized controllers, drawing inspiration from Model Predictive Control (MPC)
The controller resembles a Quadratic Programming (QP) solver of a linear MPC problem, with the parameters of the controller being trained via Deep Reinforcement Learning (DRL)
The proposed controller is significantly more computationally efficient compared to MPC and requires fewer parameters to learn than controllers.
arXiv Detail & Related papers (2023-12-08T19:33:22Z) - A Dynamic Feedforward Control Strategy for Energy-efficient Building
System Operation [59.56144813928478]
In current control strategies and optimization algorithms, most of them rely on receiving information from real-time feedback.
We propose an engineer-friendly control strategy framework that embeds dynamic prior knowledge from building system characteristics simultaneously for system control.
We tested it in a case for heating system control with typical control strategies, which shows our framework owns a further energy-saving potential of 15%.
arXiv Detail & Related papers (2023-01-23T09:07:07Z) - Steady-State Error Compensation in Reference Tracking and Disturbance
Rejection Problems for Reinforcement Learning-Based Control [0.9023847175654602]
Reinforcement learning (RL) is a promising, upcoming topic in automatic control applications.
Initiative action state augmentation (IASA) for actor-critic-based RL controllers is introduced.
This augmentation does not require any expert knowledge, leaving the approach model free.
arXiv Detail & Related papers (2022-01-31T16:29:19Z) - Policy Search for Model Predictive Control with Application to Agile
Drone Flight [56.24908013905407]
We propose a policy-search-for-model-predictive-control framework for MPC.
Specifically, we formulate the MPC as a parameterized controller, where the hard-to-optimize decision variables are represented as high-level policies.
Experiments show that our controller achieves robust and real-time control performance in both simulation and the real world.
arXiv Detail & Related papers (2021-12-07T17:39:24Z) - Optimization of the Model Predictive Control Meta-Parameters Through
Reinforcement Learning [1.4069478981641936]
We propose a novel framework in which any parameter of the control algorithm can be jointly tuned using reinforcement learning(RL)
We demonstrate our framework on the inverted pendulum control task, reducing the total time of the control system by 36% while also improving the control performance by 18.4% over the best-performing MPC baseline.
arXiv Detail & Related papers (2021-11-07T18:33:22Z) - Evaluating model-based planning and planner amortization for continuous
control [79.49319308600228]
We take a hybrid approach, combining model predictive control (MPC) with a learned model and model-free policy learning.
We find that well-tuned model-free agents are strong baselines even for high DoF control problems.
We show that it is possible to distil a model-based planner into a policy that amortizes the planning without any loss of performance.
arXiv Detail & Related papers (2021-10-07T12:00:40Z) - Regret-optimal Estimation and Control [52.28457815067461]
We show that the regret-optimal estimator and regret-optimal controller can be derived in state-space form.
We propose regret-optimal analogs of Model-Predictive Control (MPC) and the Extended KalmanFilter (EKF) for systems with nonlinear dynamics.
arXiv Detail & Related papers (2021-06-22T23:14:21Z) - Learning a Contact-Adaptive Controller for Robust, Efficient Legged
Locomotion [95.1825179206694]
We present a framework that synthesizes robust controllers for a quadruped robot.
A high-level controller learns to choose from a set of primitives in response to changes in the environment.
A low-level controller that utilizes an established control method to robustly execute the primitives.
arXiv Detail & Related papers (2020-09-21T16:49:26Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.