Robust Neural Posterior Estimation and Statistical Model Criticism
- URL: http://arxiv.org/abs/2210.06564v1
- Date: Wed, 12 Oct 2022 20:06:55 GMT
- Title: Robust Neural Posterior Estimation and Statistical Model Criticism
- Authors: Daniel Ward, Patrick Cannon, Mark Beaumont, Matteo Fasiolo, Sebastian
M Schmon
- Abstract summary: We argue that modellers must treat simulators as idealistic representations of the true data generating process.
In this work we revisit neural posterior estimation (NPE), a class of algorithms that enable black-box parameter inference in simulation models.
We find that the presence of misspecification, in contrast, leads to unreliable inference when NPE is used naively.
- Score: 1.5749416770494706
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Computer simulations have proven a valuable tool for understanding complex
phenomena across the sciences. However, the utility of simulators for modelling
and forecasting purposes is often restricted by low data quality, as well as
practical limits to model fidelity. In order to circumvent these difficulties,
we argue that modellers must treat simulators as idealistic representations of
the true data generating process, and consequently should thoughtfully consider
the risk of model misspecification. In this work we revisit neural posterior
estimation (NPE), a class of algorithms that enable black-box parameter
inference in simulation models, and consider the implication of a
simulation-to-reality gap. While recent works have demonstrated reliable
performance of these methods, the analyses have been performed using synthetic
data generated by the simulator model itself, and have therefore only addressed
the well-specified case. In this paper, we find that the presence of
misspecification, in contrast, leads to unreliable inference when NPE is used
naively. As a remedy we argue that principled scientific inquiry with
simulators should incorporate a model criticism component, to facilitate
interpretable identification of misspecification and a robust inference
component, to fit 'wrong but useful' models. We propose robust neural posterior
estimation (RNPE), an extension of NPE to simultaneously achieve both these
aims, through explicitly modelling the discrepancies between simulations and
the observed data. We assess the approach on a range of artificially
misspecified examples, and find RNPE performs well across the tasks, whereas
naively using NPE leads to misleading and erratic posteriors.
Related papers
- All-in-one simulation-based inference [19.41881319338419]
We present a new amortized inference method -- the Simformer -- which overcomes current limitations.
The Simformer outperforms current state-of-the-art amortized inference approaches on benchmark tasks.
It can be applied to models with function-valued parameters, it can handle inference scenarios with missing or unstructured data, and it can sample arbitrary conditionals of the joint distribution of parameters and data.
arXiv Detail & Related papers (2024-04-15T10:12:33Z) - Diffusion posterior sampling for simulation-based inference in tall data settings [53.17563688225137]
Simulation-based inference ( SBI) is capable of approximating the posterior distribution that relates input parameters to a given observation.
In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model.
We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.
arXiv Detail & Related papers (2024-04-11T09:23:36Z) - Towards Theoretical Understandings of Self-Consuming Generative Models [56.84592466204185]
This paper tackles the emerging challenge of training generative models within a self-consuming loop.
We construct a theoretical framework to rigorously evaluate how this training procedure impacts the data distributions learned by future models.
We present results for kernel density estimation, delivering nuanced insights such as the impact of mixed data training on error propagation.
arXiv Detail & Related papers (2024-02-19T02:08:09Z) - A PAC-Bayesian Perspective on the Interpolating Information Criterion [54.548058449535155]
We show how a PAC-Bayes bound is obtained for a general class of models, characterizing factors which influence performance in the interpolating regime.
We quantify how the test error for overparameterized models achieving effectively zero training error depends on the quality of the implicit regularization imposed by e.g. the combination of model, parameter-initialization scheme.
arXiv Detail & Related papers (2023-11-13T01:48:08Z) - Misspecification-robust Sequential Neural Likelihood for
Simulation-based Inference [0.20971479389679337]
We propose a novel SNL method, which through the incorporation of additional adjustment parameters, is robust to model misspecification.
We demonstrate the efficacy of our approach through several illustrative examples.
arXiv Detail & Related papers (2023-01-31T02:28:18Z) - Neural Posterior Estimation with Differentiable Simulators [58.720142291102135]
We present a new method to perform Neural Posterior Estimation (NPE) with a differentiable simulator.
We demonstrate how gradient information helps constrain the shape of the posterior and improves sample-efficiency.
arXiv Detail & Related papers (2022-07-12T16:08:04Z) - Nonparametric likelihood-free inference with Jensen-Shannon divergence
for simulator-based models with categorical output [1.4298334143083322]
Likelihood-free inference for simulator-based statistical models has attracted a surge of interest, both in the machine learning and statistics communities.
Here we derive a set of theoretical results to enable estimation, hypothesis testing and construction of confidence intervals for model parameters using computation properties of the Jensen-Shannon- divergence.
Such approximation offers a rapid alternative to more-intensive approaches and can be attractive for diverse applications of simulator-based models.
arXiv Detail & Related papers (2022-05-22T18:00:13Z) - Mixed Effects Neural ODE: A Variational Approximation for Analyzing the
Dynamics of Panel Data [50.23363975709122]
We propose a probabilistic model called ME-NODE to incorporate (fixed + random) mixed effects for analyzing panel data.
We show that our model can be derived using smooth approximations of SDEs provided by the Wong-Zakai theorem.
We then derive Evidence Based Lower Bounds for ME-NODE, and develop (efficient) training algorithms.
arXiv Detail & Related papers (2022-02-18T22:41:51Z) - Black-box Bayesian inference for economic agent-based models [0.0]
We investigate the efficacy of two classes of black-box approximate Bayesian inference methods.
We demonstrate that neural network based black-box methods provide state of the art parameter inference for economic simulation models.
arXiv Detail & Related papers (2022-02-01T18:16:12Z) - Deep Bayesian Active Learning for Accelerating Stochastic Simulation [74.58219903138301]
Interactive Neural Process (INP) is a deep active learning framework for simulations and with active learning approaches.
For active learning, we propose a novel acquisition function, Latent Information Gain (LIG), calculated in the latent space of NP based models.
The results demonstrate STNP outperforms the baselines in the learning setting and LIG achieves the state-of-the-art for active learning.
arXiv Detail & Related papers (2021-06-05T01:31:51Z) - Amortized Bayesian model comparison with evidential deep learning [0.12314765641075436]
We propose a novel method for performing Bayesian model comparison using specialized deep learning architectures.
Our method is purely simulation-based and circumvents the step of explicitly fitting all alternative models under consideration to each observed dataset.
We show that our method achieves excellent results in terms of accuracy, calibration, and efficiency across the examples considered in this work.
arXiv Detail & Related papers (2020-04-22T15:15:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.