Consistency of Empirical Bayes And Kernel Flow For Hierarchical
Parameter Estimation
- URL: http://arxiv.org/abs/2005.11375v2
- Date: Tue, 16 Mar 2021 23:57:46 GMT
- Title: Consistency of Empirical Bayes And Kernel Flow For Hierarchical
Parameter Estimation
- Authors: Yifan Chen, Houman Owhadi, Andrew M. Stuart
- Abstract summary: This paper studies two paradigms of learning hierarchical parameters.
One is from the probabilistic Bayesian perspective, in particular, the empirical Bayes approach.
The other is from the deterministic and approximation theoretic view, and in particular the kernel flow algorithm.
- Score: 18.852474191260445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Gaussian process regression has proven very powerful in statistics, machine
learning and inverse problems. A crucial aspect of the success of this
methodology, in a wide range of applications to complex and real-world
problems, is hierarchical modeling and learning of hyperparameters. The purpose
of this paper is to study two paradigms of learning hierarchical parameters:
one is from the probabilistic Bayesian perspective, in particular, the
empirical Bayes approach that has been largely used in Bayesian statistics; the
other is from the deterministic and approximation theoretic view, and in
particular the kernel flow algorithm that was proposed recently in the machine
learning literature. Analysis of their consistency in the large data limit, as
well as explicit identification of their implicit bias in parameter learning,
are established in this paper for a Mat\'ern-like model on the torus. A
particular technical challenge we overcome is the learning of the regularity
parameter in the Mat\'ern-like field, for which consistency results have been
very scarce in the spatial statistics literature. Moreover, we conduct
extensive numerical experiments beyond the Mat\'ern-like model, comparing the
two algorithms further. These experiments demonstrate learning of other
hierarchical parameters, such as amplitude and lengthscale; they also
illustrate the setting of model misspecification in which the kernel flow
approach could show superior performance to the more traditional empirical
Bayes approach.
Related papers
- A sparse PAC-Bayesian approach for high-dimensional quantile prediction [0.0]
This paper presents a novel probabilistic machine learning approach for high-dimensional quantile prediction.
It uses a pseudo-Bayesian framework with a scaled Student-t prior and Langevin Monte Carlo for efficient computation.
Its effectiveness is validated through simulations and real-world data, where it performs competitively against established frequentist and Bayesian techniques.
arXiv Detail & Related papers (2024-09-03T08:01:01Z) - Diffusion posterior sampling for simulation-based inference in tall data settings [53.17563688225137]
Simulation-based inference ( SBI) is capable of approximating the posterior distribution that relates input parameters to a given observation.
In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model.
We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.
arXiv Detail & Related papers (2024-04-11T09:23:36Z) - A PAC-Bayesian Perspective on the Interpolating Information Criterion [54.548058449535155]
We show how a PAC-Bayes bound is obtained for a general class of models, characterizing factors which influence performance in the interpolating regime.
We quantify how the test error for overparameterized models achieving effectively zero training error depends on the quality of the implicit regularization imposed by e.g. the combination of model, parameter-initialization scheme.
arXiv Detail & Related papers (2023-11-13T01:48:08Z) - Simulation-Based Prior Knowledge Elicitation for Parametric Bayesian Models [2.9172603864294024]
We focus on translating domain expert knowledge into corresponding prior distributions over model parameters, a process known as prior elicitation.
A major challenge for existing elicitation methods is how to effectively utilize all of these different formats in order to formulate prior distributions that align with the expert's expectations, regardless of the model structure.
Our results support the claim that our method is largely independent of the underlying model structure and adaptable to various elicitation techniques, including quantile-based, moment-based, and histogram-based methods.
arXiv Detail & Related papers (2023-08-22T10:43:05Z) - MARS: Meta-Learning as Score Matching in the Function Space [79.73213540203389]
We present a novel approach to extracting inductive biases from a set of related datasets.
We use functional Bayesian neural network inference, which views the prior as a process and performs inference in the function space.
Our approach can seamlessly acquire and represent complex prior knowledge by metalearning the score function of the data-generating process.
arXiv Detail & Related papers (2022-10-24T15:14:26Z) - On the Influence of Enforcing Model Identifiability on Learning dynamics
of Gaussian Mixture Models [14.759688428864159]
We propose a technique for extracting submodels from singular models.
Our method enforces model identifiability during training.
We show how the method can be applied to more complex models like deep neural networks.
arXiv Detail & Related papers (2022-06-17T07:50:22Z) - A Farewell to the Bias-Variance Tradeoff? An Overview of the Theory of
Overparameterized Machine Learning [37.01683478234978]
The rapid recent progress in machine learning (ML) has raised a number of scientific questions that challenge the longstanding dogma of the field.
One of the most important riddles is the good empirical generalization of over parameterized models.
arXiv Detail & Related papers (2021-09-06T10:48:40Z) - Evaluating Sensitivity to the Stick-Breaking Prior in Bayesian
Nonparametrics [85.31247588089686]
We show that variational Bayesian methods can yield sensitivities with respect to parametric and nonparametric aspects of Bayesian models.
We provide both theoretical and empirical support for our variational approach to Bayesian sensitivity analysis.
arXiv Detail & Related papers (2021-07-08T03:40:18Z) - Leveraging Global Parameters for Flow-based Neural Posterior Estimation [90.21090932619695]
Inferring the parameters of a model based on experimental observations is central to the scientific method.
A particularly challenging setting is when the model is strongly indeterminate, i.e., when distinct sets of parameters yield identical observations.
We present a method for cracking such indeterminacy by exploiting additional information conveyed by an auxiliary set of observations sharing global parameters.
arXiv Detail & Related papers (2021-02-12T12:23:13Z) - Multiplicative noise and heavy tails in stochastic optimization [62.993432503309485]
empirical optimization is central to modern machine learning, but its role in its success is still unclear.
We show that it commonly arises in parameters of discrete multiplicative noise due to variance.
A detailed analysis is conducted in which we describe on key factors, including recent step size, and data, all exhibit similar results on state-of-the-art neural network models.
arXiv Detail & Related papers (2020-06-11T09:58:01Z) - Real-Time Model Calibration with Deep Reinforcement Learning [4.707841918805165]
We propose a novel framework for inference of model parameters based on reinforcement learning.
The proposed methodology is demonstrated and evaluated on two model-based diagnostics test cases.
arXiv Detail & Related papers (2020-06-07T00:11:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.