Episodic Gaussian Process-Based Learning Control with Vanishing Tracking
Errors
- URL: http://arxiv.org/abs/2307.04415v1
- Date: Mon, 10 Jul 2023 08:43:28 GMT
- Title: Episodic Gaussian Process-Based Learning Control with Vanishing Tracking
Errors
- Authors: Armin Lederer, Jonas Umlauft, Sandra Hirche
- Abstract summary: We develop an episodic approach for learning GP models, such that an arbitrary tracking accuracy can be guaranteed.
The effectiveness of the derived theory is demonstrated in several simulations.
- Score: 10.627020714408445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to the increasing complexity of technical systems, accurate first
principle models can often not be obtained. Supervised machine learning can
mitigate this issue by inferring models from measurement data. Gaussian process
regression is particularly well suited for this purpose due to its high
data-efficiency and its explicit uncertainty representation, which allows the
derivation of prediction error bounds. These error bounds have been exploited
to show tracking accuracy guarantees for a variety of control approaches, but
their direct dependency on the training data is generally unclear. We address
this issue by deriving a Bayesian prediction error bound for GP regression,
which we show to decay with the growth of a novel, kernel-based measure of data
density. Based on the prediction error bound, we prove time-varying tracking
accuracy guarantees for learned GP models used as feedback compensation of
unknown nonlinearities, and show to achieve vanishing tracking error with
increasing data density. This enables us to develop an episodic approach for
learning Gaussian process models, such that an arbitrary tracking accuracy can
be guaranteed. The effectiveness of the derived theory is demonstrated in
several simulations.
Related papers
- A note on the error analysis of data-driven closure models for large eddy simulations of turbulence [2.4548283109365436]
We provide a mathematical formulation for error propagation in flow trajectory prediction using data-driven turbulence closure modeling.
We retrieve an upper bound for the prediction error when utilizing a data-driven closure model.
Our analysis also shows that the error propagates exponentially with rollout time and the upper bound of the system Jacobian.
arXiv Detail & Related papers (2024-05-27T19:20:22Z) - Meta-learning to Calibrate Gaussian Processes with Deep Kernels for
Regression Uncertainty Estimation [43.23399636191726]
We propose a meta-learning method for calibrating deep kernel GPs for improving regression uncertainty estimation performance.
The proposed method meta-learns how to calibrate uncertainty using data from various tasks by minimizing the test expected calibration error.
Our experiments demonstrate that the proposed method improves uncertainty estimation performance while keeping high regression performance.
arXiv Detail & Related papers (2023-12-13T07:58:47Z) - Neural Operator Variational Inference based on Regularized Stein
Discrepancy for Deep Gaussian Processes [23.87733307119697]
We introduce Neural Operator Variational Inference (NOVI) for Deep Gaussian Processes.
NOVI uses a neural generator to obtain a sampler and minimizes the Regularized Stein Discrepancy in L2 space between the generated distribution and true posterior.
We demonstrate that the bias introduced by our method can be controlled by multiplying the divergence with a constant, which leads to robust error control and ensures the stability and precision of the algorithm.
arXiv Detail & Related papers (2023-09-22T06:56:35Z) - Can Learning Deteriorate Control? Analyzing Computational Delays in
Gaussian Process-Based Event-Triggered Online Learning [7.697964930378468]
We propose a novel event trigger for GP-based online learning with computational delays.
We show to offer advantages over offline trained GP models for sufficiently small computation times.
arXiv Detail & Related papers (2023-05-14T14:37:33Z) - Learning Sample Difficulty from Pre-trained Models for Reliable
Prediction [55.77136037458667]
We propose to utilize large-scale pre-trained models to guide downstream model training with sample difficulty-aware entropy regularization.
We simultaneously improve accuracy and uncertainty calibration across challenging benchmarks.
arXiv Detail & Related papers (2023-04-20T07:29:23Z) - Imputation-Free Learning from Incomplete Observations [73.15386629370111]
We introduce the importance of guided gradient descent (IGSGD) method to train inference from inputs containing missing values without imputation.
We employ reinforcement learning (RL) to adjust the gradients used to train the models via back-propagation.
Our imputation-free predictions outperform the traditional two-step imputation-based predictions using state-of-the-art imputation methods.
arXiv Detail & Related papers (2021-07-05T12:44:39Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z) - Improving Uncertainty Calibration via Prior Augmented Data [56.88185136509654]
Neural networks have proven successful at learning from complex data distributions by acting as universal function approximators.
They are often overconfident in their predictions, which leads to inaccurate and miscalibrated probabilistic predictions.
We propose a solution by seeking out regions of feature space where the model is unjustifiably overconfident, and conditionally raising the entropy of those predictions towards that of the prior distribution of the labels.
arXiv Detail & Related papers (2021-02-22T07:02:37Z) - Uniform Error and Posterior Variance Bounds for Gaussian Process
Regression with Application to Safe Control [11.42419064387567]
We present a novel uniform error bound using Lipschitz and an analysis of the posterior variance function for a large class of kernels.
We show how these results can be used to guarantee safe control of an unknown dynamical system.
arXiv Detail & Related papers (2021-01-13T20:06:30Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z) - Evaluating Prediction-Time Batch Normalization for Robustness under
Covariate Shift [81.74795324629712]
We call prediction-time batch normalization, which significantly improves model accuracy and calibration under covariate shift.
We show that prediction-time batch normalization provides complementary benefits to existing state-of-the-art approaches for improving robustness.
The method has mixed results when used alongside pre-training, and does not seem to perform as well under more natural types of dataset shift.
arXiv Detail & Related papers (2020-06-19T05:08:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.