Promises of Deep Kernel Learning for Control Synthesis
- URL: http://arxiv.org/abs/2309.06569v2
- Date: Wed, 13 Mar 2024 00:15:08 GMT
- Title: Promises of Deep Kernel Learning for Control Synthesis
- Authors: Robert Reed, Luca Laurenti, Morteza Lahijanian
- Abstract summary: Deep Learning (DKL) combines the representational power of neural networks with the uncertainty of Gaussian Processes.
We develop a scalable abstraction-based framework that enables the use of DKL for control synthesis of dynamical systems.
- Score: 14.401542690028554
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep Kernel Learning (DKL) combines the representational power of neural
networks with the uncertainty quantification of Gaussian Processes. Hence, it
is potentially a promising tool to learn and control complex dynamical systems.
In this work, we develop a scalable abstraction-based framework that enables
the use of DKL for control synthesis of stochastic dynamical systems against
complex specifications. Specifically, we consider temporal logic specifications
and create an end-to-end framework that uses DKL to learn an unknown system
from data and formally abstracts the DKL model into an Interval Markov Decision
Process (IMDP) to perform control synthesis with correctness guarantees.
Furthermore, we identify a deep architecture that enables accurate learning and
efficient abstraction computation. The effectiveness of our approach is
illustrated on various benchmarks, including a 5-D nonlinear stochastic system,
showing how control synthesis with DKL can substantially outperform
state-of-the-art competitive methods.
Related papers
- Controlling dynamics of stochastic systems with deep reinforcement learning [0.0]
We propose a simulation algorithm that allows achieving control of the dynamics of systems through the use of trained artificial neural networks.
Specifically, we use agent-based simulations where the neural network plays the role of the controller that drives local state-to-state transitions.
arXiv Detail & Related papers (2025-02-25T11:28:12Z) - End-to-End Learning Framework for Solving Non-Markovian Optimal Control [9.156265463755807]
We propose an innovative system identification method control strategy for FOLTI systems.
We also develop the first end-to-end data-driven learning framework, Fractional-Order Learning for Optimal Control (FOLOC)
arXiv Detail & Related papers (2025-02-07T04:18:56Z) - Dropout MPC: An Ensemble Neural MPC Approach for Systems with Learned Dynamics [0.0]
We propose a novel sampling-based ensemble neural MPC algorithm that employs the Monte-Carlo dropout technique on the learned system model.
The method aims in general at uncertain systems with complex dynamics, where models derived from first principles are hard to infer.
arXiv Detail & Related papers (2024-06-04T17:15:25Z) - Loop Polarity Analysis to Avoid Underspecification in Deep Learning [0.0]
In this paper, we turn to loop polarity analysis as a tool for specifying the causal structure of a data-generating process.
We show how measuring the polarity of the different feedback loops that compose a system can lead to more robust inferences on the part of neural networks.
arXiv Detail & Related papers (2023-09-18T23:49:42Z) - Unifying Synergies between Self-supervised Learning and Dynamic
Computation [53.66628188936682]
We present a novel perspective on the interplay between SSL and DC paradigms.
We show that it is feasible to simultaneously learn a dense and gated sub-network from scratch in a SSL setting.
The co-evolution during pre-training of both dense and gated encoder offers a good accuracy-efficiency trade-off.
arXiv Detail & Related papers (2023-01-22T17:12:58Z) - Formal Controller Synthesis for Markov Jump Linear Systems with
Uncertain Dynamics [64.72260320446158]
We propose a method for synthesising controllers for Markov jump linear systems.
Our method is based on a finite-state abstraction that captures both the discrete (mode-jumping) and continuous (stochastic linear) behaviour of the MJLS.
We apply our method to multiple realistic benchmark problems, in particular, a temperature control and an aerial vehicle delivery problem.
arXiv Detail & Related papers (2022-12-01T17:36:30Z) - Discrete-Event Controller Synthesis for Autonomous Systems with
Deep-Learning Perception Components [0.0]
We present DeepDECS, a new method for the synthesis of correct-by-construction discrete-event controllers for autonomous systems.
The synthesised models correspond to controllers guaranteed to satisfy the safety, dependability and performance requirements of the autonomous system.
We use the method in simulation to synthesise controllers for mobile-robot collision mitigation and for maintaining driver attentiveness in shared-control autonomous driving.
arXiv Detail & Related papers (2022-02-07T17:10:00Z) - Safe-Critical Modular Deep Reinforcement Learning with Temporal Logic
through Gaussian Processes and Control Barrier Functions [3.5897534810405403]
Reinforcement learning (RL) is a promising approach and has limited success towards real-world applications.
In this paper, we propose a learning-based control framework consisting of several aspects.
We show such an ECBF-based modular deep RL algorithm achieves near-perfect success rates and guard safety with a high probability.
arXiv Detail & Related papers (2021-09-07T00:51:12Z) - Probabilistic robust linear quadratic regulators with Gaussian processes [73.0364959221845]
Probabilistic models such as Gaussian processes (GPs) are powerful tools to learn unknown dynamical systems from data for subsequent use in control design.
We present a novel controller synthesis for linearized GP dynamics that yields robust controllers with respect to a probabilistic stability margin.
arXiv Detail & Related papers (2021-05-17T08:36:18Z) - Online Reinforcement Learning Control by Direct Heuristic Dynamic
Programming: from Time-Driven to Event-Driven [80.94390916562179]
Time-driven learning refers to the machine learning method that updates parameters in a prediction model continuously as new data arrives.
It is desirable to prevent the time-driven dHDP from updating due to insignificant system event such as noise.
We show how the event-driven dHDP algorithm works in comparison to the original time-driven dHDP.
arXiv Detail & Related papers (2020-06-16T05:51:25Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z) - Certified Reinforcement Learning with Logic Guidance [78.2286146954051]
We propose a model-free RL algorithm that enables the use of Linear Temporal Logic (LTL) to formulate a goal for unknown continuous-state/action Markov Decision Processes (MDPs)
The algorithm is guaranteed to synthesise a control policy whose traces satisfy the specification with maximal probability.
arXiv Detail & Related papers (2019-02-02T20:09:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.