A Rigorous Link between Deep Ensembles and (Variational) Bayesian
Methods
- URL: http://arxiv.org/abs/2305.15027v2
- Date: Sun, 22 Oct 2023 14:44:18 GMT
- Title: A Rigorous Link between Deep Ensembles and (Variational) Bayesian
Methods
- Authors: Veit David Wild, Sahra Ghalebikesabi, Dino Sejdinovic, Jeremias
Knoblauch
- Abstract summary: We establish the first mathematically rigorous link between Bayesian, variational Bayesian, and ensemble methods.
On a technical level, our contribution amounts to a generalised variational inference through the lense of Wasserstein flows.
- Score: 14.845158804951552
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We establish the first mathematically rigorous link between Bayesian,
variational Bayesian, and ensemble methods. A key step towards this it to
reformulate the non-convex optimisation problem typically encountered in deep
learning as a convex optimisation in the space of probability measures. On a
technical level, our contribution amounts to studying generalised variational
inference through the lense of Wasserstein gradient flows. The result is a
unified theory of various seemingly disconnected approaches that are commonly
used for uncertainty quantification in deep learning -- including deep
ensembles and (variational) Bayesian methods. This offers a fresh perspective
on the reasons behind the success of deep ensembles over procedures based on
parameterised variational inference, and allows the derivation of new
ensembling schemes with convergence guarantees. We showcase this by proposing a
family of interacting deep ensembles with direct parallels to the interactions
of particle systems in thermodynamics, and use our theory to prove the
convergence of these algorithms to a well-defined global minimiser on the space
of probability measures.
Related papers
- Fully Bayesian Differential Gaussian Processes through Stochastic Differential Equations [7.439555720106548]
We propose a fully Bayesian approach that treats the kernel hyper parameters as random variables and constructs coupled differential equations (SDEs) to learn their posterior distribution and that of inducing points.
Our approach provides a time-varying, comprehensive, and realistic posterior approximation through coupling variables using SDE methods.
Our work opens up exciting research avenues for advancing Bayesian inference and offers a powerful modeling tool for continuous-time Gaussian processes.
arXiv Detail & Related papers (2024-08-12T11:41:07Z) - A Unified Theory of Stochastic Proximal Point Methods without Smoothness [52.30944052987393]
Proximal point methods have attracted considerable interest owing to their numerical stability and robustness against imperfect tuning.
This paper presents a comprehensive analysis of a broad range of variations of the proximal point method (SPPM)
arXiv Detail & Related papers (2024-05-24T21:09:19Z) - On Representations of Mean-Field Variational Inference [2.4316550366482357]
We present a framework to analyze mean field variational inference (MFVI) algorithms.
Our approach enables the MFVI problem to be represented in three different manners.
Rigorous guarantees are established to show that a time-discretized implementation of the coordinate ascent variational inference algorithm yields a gradient flow in the limit.
arXiv Detail & Related papers (2022-10-20T16:26:22Z) - Guaranteed Conservation of Momentum for Learning Particle-based Fluid
Dynamics [96.9177297872723]
We present a novel method for guaranteeing linear momentum in learned physics simulations.
We enforce conservation of momentum with a hard constraint, which we realize via antisymmetrical continuous convolutional layers.
In combination, the proposed method allows us to increase the physical accuracy of the learned simulator substantially.
arXiv Detail & Related papers (2022-10-12T09:12:59Z) - The Dynamics of Riemannian Robbins-Monro Algorithms [101.29301565229265]
We propose a family of Riemannian algorithms generalizing and extending the seminal approximation framework of Robbins and Monro.
Compared to their Euclidean counterparts, Riemannian algorithms are much less understood due to lack of a global linear structure on the manifold.
We provide a general template of almost sure convergence results that mirrors and extends the existing theory for Euclidean Robbins-Monro schemes.
arXiv Detail & Related papers (2022-06-14T12:30:11Z) - Harnessing Heterogeneity: Learning from Decomposed Feedback in Bayesian
Modeling [68.69431580852535]
We introduce a novel GP regression to incorporate the subgroup feedback.
Our modified regression has provably lower variance -- and thus a more accurate posterior -- compared to previous approaches.
We execute our algorithm on two disparate social problems.
arXiv Detail & Related papers (2021-07-07T03:57:22Z) - A Discrete Variational Derivation of Accelerated Methods in Optimization [68.8204255655161]
We introduce variational which allow us to derive different methods for optimization.
We derive two families of optimization methods in one-to-one correspondence.
The preservation of symplecticity of autonomous systems occurs here solely on the fibers.
arXiv Detail & Related papers (2021-06-04T20:21:53Z) - Acceleration Methods [57.202881673406324]
We first use quadratic optimization problems to introduce two key families of acceleration methods.
We discuss momentum methods in detail, starting with the seminal work of Nesterov.
We conclude by discussing restart schemes, a set of simple techniques for reaching nearly optimal convergence rates.
arXiv Detail & Related papers (2021-01-23T17:58:25Z) - Distributed Variational Bayesian Algorithms Over Sensor Networks [6.572330981878818]
We propose two novel distributed VB algorithms for general Bayesian inference problem.
The proposed algorithms have excellent performance, which are almost as good as the corresponding centralized VB algorithm relying on all data available in a fusion center.
arXiv Detail & Related papers (2020-11-27T08:12:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.