Hierarchical Inference With Bayesian Neural Networks: An Application to
Strong Gravitational Lensing
- URL: http://arxiv.org/abs/2010.13787v3
- Date: Mon, 22 Mar 2021 19:23:39 GMT
- Title: Hierarchical Inference With Bayesian Neural Networks: An Application to
Strong Gravitational Lensing
- Authors: Sebastian Wagner-Carena, Ji Won Park, Simon Birrer, Philip J.
Marshall, Aaron Roodman, Risa H. Wechsler (for the LSST Dark Energy Science
Collaboration)
- Abstract summary: We develop a hierarchical inference framework that mitigates the bias introduced by an unrepresentative training set's interim prior.
Our full pipeline, from training to hierarchical inference on thousands of lenses, can be run in a day.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the past few years, approximate Bayesian Neural Networks (BNNs) have
demonstrated the ability to produce statistically consistent posteriors on a
wide range of inference problems at unprecedented speed and scale. However, any
disconnect between training sets and the distribution of real-world objects can
introduce bias when BNNs are applied to data. This is a common challenge in
astrophysics and cosmology, where the unknown distribution of objects in our
Universe is often the science goal. In this work, we incorporate BNNs with
flexible posterior parameterizations into a hierarchical inference framework
that allows for the reconstruction of population hyperparameters and removes
the bias introduced by the training distribution. We focus on the challenge of
producing posterior PDFs for strong gravitational lens mass model parameters
given Hubble Space Telescope (HST) quality single-filter, lens-subtracted,
synthetic imaging data. We show that the posterior PDFs are sufficiently
accurate (i.e., statistically consistent with the truth) across a wide variety
of power-law elliptical lens mass distributions. We then apply our approach to
test data sets whose lens parameters are drawn from distributions that are
drastically different from the training set. We show that our hierarchical
inference framework mitigates the bias introduced by an unrepresentative
training set's interim prior. Simultaneously, given a sufficiently broad
training set, we can precisely reconstruct the population hyperparameters
governing our test distributions. Our full pipeline, from training to
hierarchical inference on thousands of lenses, can be run in a day. The
framework presented here will allow us to efficiently exploit the full
constraining power of future ground- and space-based surveys.
Related papers
- Amortizing intractable inference in diffusion models for vision, language, and control [89.65631572949702]
This paper studies amortized sampling of the posterior over data, $mathbfxsim prm post(mathbfx)propto p(mathbfx)r(mathbfx)$, in a model that consists of a diffusion generative model prior $p(mathbfx)$ and a black-box constraint or function $r(mathbfx)$.
We prove the correctness of a data-free learning objective, relative trajectory balance, for training a diffusion model that samples from
arXiv Detail & Related papers (2024-05-31T16:18:46Z) - Sparsifying Bayesian neural networks with latent binary variables and
normalizing flows [10.865434331546126]
We will consider two extensions to the latent binary Bayesian neural networks (LBBNN) method.
Firstly, by using the local reparametrization trick (LRT) to sample the hidden units directly, we get a more computationally efficient algorithm.
More importantly, by using normalizing flows on the variational posterior distribution of the LBBNN parameters, the network learns a more flexible variational posterior distribution than the mean field Gaussian.
arXiv Detail & Related papers (2023-05-05T09:40:28Z) - Distributionally Robust Models with Parametric Likelihood Ratios [123.05074253513935]
Three simple ideas allow us to train models with DRO using a broader class of parametric likelihood ratios.
We find that models trained with the resulting parametric adversaries are consistently more robust to subpopulation shifts when compared to other DRO approaches.
arXiv Detail & Related papers (2022-04-13T12:43:12Z) - Estimating permeability of 3D micro-CT images by physics-informed CNNs
based on DNS [1.6274397329511197]
This paper presents a novel methodology for permeability prediction from micro-CT scans of geological rock samples.
The training data set for CNNs dedicated to permeability prediction consists of permeability labels that are typically generated by classical lattice Boltzmann methods (LBM)
We instead perform direct numerical simulation (DNS) by solving the stationary Stokes equation in an efficient and distributed-parallel manner.
arXiv Detail & Related papers (2021-09-04T08:43:19Z) - Bayesian Deep Learning for Partial Differential Equation Parameter
Discovery with Sparse and Noisy Data [0.0]
We propose to use Bayesian neural networks (BNN) in order to recover the full system states from measurement data.
We show that it is possible to accurately capture physics of varying complexity without overfitting.
We demonstrate our approach on a handful of example applied to physics and non-linear dynamics.
arXiv Detail & Related papers (2021-08-05T19:43:15Z) - Attribute-Guided Adversarial Training for Robustness to Natural
Perturbations [64.35805267250682]
We propose an adversarial training approach which learns to generate new samples so as to maximize exposure of the classifier to the attributes-space.
Our approach enables deep neural networks to be robust against a wide range of naturally occurring perturbations.
arXiv Detail & Related papers (2020-12-03T10:17:30Z) - Large-Scale Gravitational Lens Modeling with Bayesian Neural Networks
for Accurate and Precise Inference of the Hubble Constant [0.0]
We investigate the use of approximate Bayesian neural networks (BNNs) in modeling hundreds of time-delay gravitational lenses.
A simple combination of 200 test-set lenses results in a precision of 0.5 $textrmkm s-1 textrm Mpc-1$ ($0.7%$)
Our pipeline is a promising tool for exploring ensemble-level systematics in lens modeling.
arXiv Detail & Related papers (2020-11-30T19:00:20Z) - Lessons Learned from the Training of GANs on Artificial Datasets [0.0]
Generative Adversarial Networks (GANs) have made great progress in synthesizing realistic images in recent years.
GANs are prone to underfitting or overfitting, making the analysis of them difficult and constrained.
We train them on artificial datasets where there are infinitely many samples and the real data distributions are simple.
We find that training mixtures of GANs leads to more performance gain compared to increasing the network depth or width.
arXiv Detail & Related papers (2020-07-13T14:51:02Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z) - GANs with Conditional Independence Graphs: On Subadditivity of
Probability Divergences [70.30467057209405]
Generative Adversarial Networks (GANs) are modern methods to learn the underlying distribution of a data set.
GANs are designed in a model-free fashion where no additional information about the underlying distribution is available.
We propose a principled design of a model-based GAN that uses a set of simple discriminators on the neighborhoods of the Bayes-net/MRF.
arXiv Detail & Related papers (2020-03-02T04:31:22Z) - When Relation Networks meet GANs: Relation GANs with Triplet Loss [110.7572918636599]
Training stability is still a lingering concern of generative adversarial networks (GANs)
In this paper, we explore a relation network architecture for the discriminator and design a triplet loss which performs better generalization and stability.
Experiments on benchmark datasets show that the proposed relation discriminator and new loss can provide significant improvement on variable vision tasks.
arXiv Detail & Related papers (2020-02-24T11:35:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.