Physics Inspired Approaches To Understanding Gaussian Processes
- URL: http://arxiv.org/abs/2305.10748v2
- Date: Tue, 6 Jun 2023 16:52:52 GMT
- Title: Physics Inspired Approaches To Understanding Gaussian Processes
- Authors: Maximilian P. Niroomand and Luke Dicks and Edward O. Pyzer-Knapp and
David J. Wales
- Abstract summary: We contribute an analysis of the loss landscape for GP models using methods from physics.
We demonstrate $nu$-continuity for Matern kernels and outline aspects of catastrophe theory at critical points in the loss landscape.
We also provide an a priori method for evaluating the effect of GP ensembles and discuss various voting approaches based on physical properties of the loss landscape.
- Score: 0.9712140341805067
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Prior beliefs about the latent function to shape inductive biases can be
incorporated into a Gaussian Process (GP) via the kernel. However, beyond
kernel choices, the decision-making process of GP models remains poorly
understood. In this work, we contribute an analysis of the loss landscape for
GP models using methods from physics. We demonstrate $\nu$-continuity for
Matern kernels and outline aspects of catastrophe theory at critical points in
the loss landscape. By directly including $\nu$ in the hyperparameter
optimisation for Matern kernels, we find that typical values of $\nu$ are far
from optimal in terms of performance, yet prevail in the literature due to the
increased computational speed. We also provide an a priori method for
evaluating the effect of GP ensembles and discuss various voting approaches
based on physical properties of the loss landscape. The utility of these
approaches is demonstrated for various synthetic and real datasets. Our
findings provide an enhanced understanding of the decision-making process
behind GPs and offer practical guidance for improving their performance and
interpretability in a range of applications.
Related papers
- Kernel Approximation of Fisher-Rao Gradient Flows [52.154685604660465]
We present a rigorous investigation of Fisher-Rao and Wasserstein type gradient flows concerning their gradient structures, flow equations, and their kernel approximations.
Specifically, we focus on the Fisher-Rao geometry and its various kernel-based approximations, developing a principled theoretical framework.
arXiv Detail & Related papers (2024-10-27T22:52:08Z) - Amortized Variational Inference for Deep Gaussian Processes [0.0]
Deep Gaussian processes (DGPs) are multilayer generalizations of Gaussian processes (GPs)
We introduce amortized variational inference for DGPs, which learns an inference function that maps each observation to variational parameters.
Our method performs similarly or better than previous approaches at less computational cost.
arXiv Detail & Related papers (2024-09-18T20:23:27Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - FaDIn: Fast Discretized Inference for Hawkes Processes with General
Parametric Kernels [82.53569355337586]
This work offers an efficient solution to temporal point processes inference using general parametric kernels with finite support.
The method's effectiveness is evaluated by modeling the occurrence of stimuli-induced patterns from brain signals recorded with magnetoencephalography (MEG)
Results show that the proposed approach leads to an improved estimation of pattern latency than the state-of-the-art.
arXiv Detail & Related papers (2022-10-10T12:35:02Z) - Meta-learning Feature Representations for Adaptive Gaussian Processes
via Implicit Differentiation [1.5293427903448025]
We propose a general framework for learning deep kernels by interpolating between meta-learning and conventional learning.
Although ADKF is a completely general method, we argue that it is especially well-suited for drug discovery problems.
arXiv Detail & Related papers (2022-05-05T15:26:53Z) - Robust and Adaptive Temporal-Difference Learning Using An Ensemble of
Gaussian Processes [70.80716221080118]
The paper takes a generative perspective on policy evaluation via temporal-difference (TD) learning.
The OS-GPTD approach is developed to estimate the value function for a given policy by observing a sequence of state-reward pairs.
To alleviate the limited expressiveness associated with a single fixed kernel, a weighted ensemble (E) of GP priors is employed to yield an alternative scheme.
arXiv Detail & Related papers (2021-12-01T23:15:09Z) - Incremental Ensemble Gaussian Processes [53.3291389385672]
We propose an incremental ensemble (IE-) GP framework, where an EGP meta-learner employs an it ensemble of GP learners, each having a unique kernel belonging to a prescribed kernel dictionary.
With each GP expert leveraging the random feature-based approximation to perform online prediction and model update with it scalability, the EGP meta-learner capitalizes on data-adaptive weights to synthesize the per-expert predictions.
The novel IE-GP is generalized to accommodate time-varying functions by modeling structured dynamics at the EGP meta-learner and within each GP learner.
arXiv Detail & Related papers (2021-10-13T15:11:25Z) - Adversarial Robustness Guarantees for Gaussian Processes [22.403365399119107]
Gaussian processes (GPs) enable principled computation of model uncertainty, making them attractive for safety-critical applications.
We present a framework to analyse adversarial robustness of GPs, defined as invariance of the model's decision to bounded perturbations.
We develop a branch-and-bound scheme to refine the bounds and show, for any $epsilon > 0$, that our algorithm is guaranteed to converge to values $epsilon$-close to the actual values in finitely many iterations.
arXiv Detail & Related papers (2021-04-07T15:14:56Z) - KrigHedge: Gaussian Process Surrogates for Delta Hedging [0.0]
We investigate a machine learning approach to option Greeks approximation based on Gaussian process (GP) surrogates.
We provide a detailed analysis of numerous aspects of GP surrogates, including choice of kernel family, simulation design, choice of trend function and impact of noise.
We discuss the application to Delta hedging, including a new Lemma that relates quality of the Delta approximation to discrete-time hedging loss.
arXiv Detail & Related papers (2020-10-16T14:08:13Z) - Beyond variance reduction: Understanding the true impact of baselines on
policy optimization [24.09670734037029]
We show that learning dynamics are governed by the curvature of the loss function and the noise of the gradient estimates.
We present theoretical results showing that, at least for bandit problems, curvature and noise are not sufficient to explain the learning dynamics.
arXiv Detail & Related papers (2020-08-31T17:52:09Z) - SLEIPNIR: Deterministic and Provably Accurate Feature Expansion for
Gaussian Process Regression with Derivatives [86.01677297601624]
We propose a novel approach for scaling GP regression with derivatives based on quadrature Fourier features.
We prove deterministic, non-asymptotic and exponentially fast decaying error bounds which apply for both the approximated kernel as well as the approximated posterior.
arXiv Detail & Related papers (2020-03-05T14:33:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.