Neural Lyapunov Control for Discrete-Time Systems
- URL: http://arxiv.org/abs/2305.06547v3
- Date: Sun, 24 Dec 2023 20:50:33 GMT
- Title: Neural Lyapunov Control for Discrete-Time Systems
- Authors: Junlin Wu, Andrew Clark, Yiannis Kantaros and Yevgeniy Vorobeychik
- Abstract summary: A general approach is to compute a combination of a Lyapunov function and an associated control policy.
Several methods have been proposed that represent Lyapunov functions using neural networks.
We propose the first approach for learning neural Lyapunov control in a broad class of discrete-time systems.
- Score: 30.135651803114307
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While ensuring stability for linear systems is well understood, it remains a
major challenge for nonlinear systems. A general approach in such cases is to
compute a combination of a Lyapunov function and an associated control policy.
However, finding Lyapunov functions for general nonlinear systems is a
challenging task. To address this challenge, several methods have been proposed
that represent Lyapunov functions using neural networks. However, such
approaches either focus on continuous-time systems, or highly restricted
classes of nonlinear dynamics. We propose the first approach for learning
neural Lyapunov control in a broad class of discrete-time systems. Three key
ingredients enable us to effectively learn provably stable control policies.
The first is a novel mixed-integer linear programming approach for verifying
the discrete-time Lyapunov stability conditions, leveraging the particular
structure of these conditions. The second is a novel approach for computing
verified sublevel sets. The third is a heuristic gradient-based method for
quickly finding counterexamples to significantly speed up Lyapunov function
learning. Our experiments on four standard benchmarks demonstrate that our
approach significantly outperforms state-of-the-art baselines. For example, on
the path tracking benchmark, we outperform recent neural Lyapunov control
baselines by an order of magnitude in both running time and the size of the
region of attraction, and on two of the four benchmarks (cartpole and PVTOL),
ours is the first automated approach to return a provably stable controller.
Our code is available at: https://github.com/jlwu002/nlc_discrete.
Related papers
- Lyapunov-stable Neural Control for State and Output Feedback: A Novel Formulation [67.63756749551924]
Learning-based neural network (NN) control policies have shown impressive empirical performance in a wide range of tasks in robotics and control.
Lyapunov stability guarantees over the region-of-attraction (ROA) for NN controllers with nonlinear dynamical systems are challenging to obtain.
We demonstrate a new framework for learning NN controllers together with Lyapunov certificates using fast empirical falsification and strategic regularizations.
arXiv Detail & Related papers (2024-04-11T17:49:15Z) - Interval Reachability of Nonlinear Dynamical Systems with Neural Network
Controllers [5.543220407902113]
This paper proposes a computationally efficient framework, based on interval analysis, for rigorous verification of nonlinear continuous-time dynamical systems with neural network controllers.
Inspired by mixed monotone theory, we embed the closed-loop dynamics into a larger system using an inclusion function of the neural network and a decomposition function of the open-loop system.
We show that one can efficiently compute hyper-rectangular over-approximations of the reachable sets using a single trajectory of the embedding system.
arXiv Detail & Related papers (2023-01-19T06:46:36Z) - Backward Reachability Analysis of Neural Feedback Loops: Techniques for
Linear and Nonlinear Systems [59.57462129637796]
This paper presents a backward reachability approach for safety verification of closed-loop systems with neural networks (NNs)
The presence of NNs in the feedback loop presents a unique set of problems due to the nonlinearities in their activation functions and because NN models are generally not invertible.
We present frameworks for calculating BP over-approximations for both linear and nonlinear systems with control policies represented by feedforward NNs.
arXiv Detail & Related papers (2022-09-28T13:17:28Z) - Neural Lyapunov Control of Unknown Nonlinear Systems with Stability
Guarantees [4.786698731084036]
We propose a learning framework to stabilize an unknown nonlinear system with a neural controller and learn a neural Lyapunov function.
We provide theoretical guarantees of the proposed learning framework in terms of the closed-loop stability for the unknown nonlinear system.
arXiv Detail & Related papers (2022-06-04T05:57:31Z) - Neural System Level Synthesis: Learning over All Stabilizing Policies
for Nonlinear Systems [0.0]
We propose a Neural SLS (Neur-SLS) approach guaranteeing closed-loop stability during and after parameter optimization.
We exploit recent Deep Neural Network (DNN) models based on Recurrent Equilibrium Networks (RENs) to learn over a rich class of nonlinear stable operators.
arXiv Detail & Related papers (2022-03-22T15:22:31Z) - Stability Verification in Stochastic Control Systems via Neural Network
Supermartingales [17.558766911646263]
We present an approach for general nonlinear control problems with two novel aspects.
We use ranking supergales (RSMs) to certify a.s.asymptotic stability, and we present a method for learning neural networks.
arXiv Detail & Related papers (2021-12-17T13:05:14Z) - Imitation Learning of Stabilizing Policies for Nonlinear Systems [1.52292571922932]
It is shown that the methods developed for linear systems and controllers can be readily extended to controllers using sum of squares.
A projected gradient descent algorithm and an alternating direction method of algorithm are proposed ass for the stabilizing imitation learning problem.
arXiv Detail & Related papers (2021-09-22T17:27:19Z) - Learning the Linear Quadratic Regulator from Nonlinear Observations [135.66883119468707]
We introduce a new problem setting for continuous control called the LQR with Rich Observations, or RichLQR.
In our setting, the environment is summarized by a low-dimensional continuous latent state with linear dynamics and quadratic costs.
Our results constitute the first provable sample complexity guarantee for continuous control with an unknown nonlinearity in the system model and general function approximation.
arXiv Detail & Related papers (2020-10-08T07:02:47Z) - Reinforcement Learning with Fast Stabilization in Linear Dynamical
Systems [91.43582419264763]
We study model-based reinforcement learning (RL) in unknown stabilizable linear dynamical systems.
We propose an algorithm that certifies fast stabilization of the underlying system by effectively exploring the environment.
We show that the proposed algorithm attains $tildemathcalO(sqrtT)$ regret after $T$ time steps of agent-environment interaction.
arXiv Detail & Related papers (2020-07-23T23:06:40Z) - Formal Synthesis of Lyapunov Neural Networks [61.79595926825511]
We propose an automatic and formally sound method for synthesising Lyapunov functions.
We employ a counterexample-guided approach where a numerical learner and a symbolic verifier interact to construct provably correct Lyapunov neural networks.
Our method synthesises Lyapunov functions faster and over wider spatial domains than the alternatives, yet providing stronger or equal guarantees.
arXiv Detail & Related papers (2020-03-19T17:21:02Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.