Neural Identification for Control
- URL: http://arxiv.org/abs/2009.11782v4
- Date: Wed, 16 Mar 2022 03:01:11 GMT
- Title: Neural Identification for Control
- Authors: Priyabrata Saha, Magnus Egerstedt, and Saibal Mukhopadhyay
- Abstract summary: The proposed method relies on the Lyapunov stability theory to generate a stable closed-loop dynamics hypothesis and corresponding control law.
We demonstrate our method on various nonlinear control problems such as n-link pendulum balancing and trajectory tracking, pendulum on cart balancing, and wheeled vehicle path following.
- Score: 30.91037635723668
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a new method for learning control law that stabilizes an unknown
nonlinear dynamical system at an equilibrium point. We formulate a system
identification task in a self-supervised learning setting that jointly learns a
controller and corresponding stable closed-loop dynamics hypothesis. The
input-output behavior of the unknown dynamical system under random control
inputs is used as the supervising signal to train the neural network-based
system model and the controller. The proposed method relies on the Lyapunov
stability theory to generate a stable closed-loop dynamics hypothesis and
corresponding control law. We demonstrate our method on various nonlinear
control problems such as n-link pendulum balancing and trajectory tracking,
pendulum on cart balancing, and wheeled vehicle path following.
Related papers
- Identification For Control Based on Neural Networks: Approximately Linearizable Models [42.15267357325546]
This work presents a control-oriented identification scheme for efficient control design and stability analysis of nonlinear systems.
Neural networks are used to identify a discrete-time nonlinear state-space model to approximate time-domain input-output behavior.
The network is constructed such that the identified model is approximately linearizable by feedback, ensuring that the control law trivially follows from the learning stage.
arXiv Detail & Related papers (2024-09-24T08:31:22Z) - Data-Driven Control with Inherent Lyapunov Stability [3.695480271934742]
We propose Control with Inherent Lyapunov Stability (CoILS) as a method for jointly learning parametric representations of a nonlinear dynamics model and a stabilizing controller from data.
In addition to the stabilizability of the learned dynamics guaranteed by our novel construction, we show that the learned controller stabilizes the true dynamics under certain assumptions on the fidelity of the learned dynamics.
arXiv Detail & Related papers (2023-03-06T14:21:42Z) - Learning Control-Oriented Dynamical Structure from Data [25.316358215670274]
We discuss a state-dependent nonlinear tracking controller formulation for general nonlinear control-affine systems.
We empirically demonstrate the efficacy of learned versions of this controller in stable trajectory tracking.
arXiv Detail & Related papers (2023-02-06T02:01:38Z) - Learning Stabilizable Deep Dynamics Models [1.75320459412718]
We propose a new method for learning the dynamics of input-affine control systems.
An important feature is that a stabilizing controller and control Lyapunov function of the learned model are obtained as well.
The proposed method can also be applied to solving Hamilton-Jacobi inequalities.
arXiv Detail & Related papers (2022-03-18T03:09:24Z) - Neural Koopman Lyapunov Control [0.0]
We propose a framework to identify and construct stabilizable bilinear control systems and its associated observables from data.
Our proposed approach provides provable guarantees of global stability for the nonlinear control systems with unknown dynamics.
arXiv Detail & Related papers (2022-01-13T17:38:09Z) - The Impact of Data on the Stability of Learning-Based Control- Extended
Version [63.97366815968177]
We propose a Lyapunov-based measure for quantifying the impact of data on the certifiable control performance.
By modeling unknown system dynamics through Gaussian processes, we can determine the interrelation between model uncertainty and satisfaction of stability conditions.
arXiv Detail & Related papers (2020-11-20T19:10:01Z) - Gaussian Process-based Min-norm Stabilizing Controller for
Control-Affine Systems with Uncertain Input Effects and Dynamics [90.81186513537777]
We propose a novel compound kernel that captures the control-affine nature of the problem.
We show that this resulting optimization problem is convex, and we call it Gaussian Process-based Control Lyapunov Function Second-Order Cone Program (GP-CLF-SOCP)
arXiv Detail & Related papers (2020-11-14T01:27:32Z) - Learning Stabilizing Controllers for Unstable Linear Quadratic
Regulators from a Single Trajectory [85.29718245299341]
We study linear controllers under quadratic costs model also known as linear quadratic regulators (LQR)
We present two different semi-definite programs (SDP) which results in a controller that stabilizes all systems within an ellipsoid uncertainty set.
We propose an efficient data dependent algorithm -- textsceXploration -- that with high probability quickly identifies a stabilizing controller.
arXiv Detail & Related papers (2020-06-19T08:58:57Z) - Active Learning for Nonlinear System Identification with Guarantees [102.43355665393067]
We study a class of nonlinear dynamical systems whose state transitions depend linearly on a known feature embedding of state-action pairs.
We propose an active learning approach that achieves this by repeating three steps: trajectory planning, trajectory tracking, and re-estimation of the system from all available data.
We show that our method estimates nonlinear dynamical systems at a parametric rate, similar to the statistical rate of standard linear regression.
arXiv Detail & Related papers (2020-06-18T04:54:11Z) - Online Reinforcement Learning Control by Direct Heuristic Dynamic
Programming: from Time-Driven to Event-Driven [80.94390916562179]
Time-driven learning refers to the machine learning method that updates parameters in a prediction model continuously as new data arrives.
It is desirable to prevent the time-driven dHDP from updating due to insignificant system event such as noise.
We show how the event-driven dHDP algorithm works in comparison to the original time-driven dHDP.
arXiv Detail & Related papers (2020-06-16T05:51:25Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.