Gravitational-wave parameter estimation with autoregressive neural
network flows
- URL: http://arxiv.org/abs/2002.07656v1
- Date: Tue, 18 Feb 2020 15:44:04 GMT
- Title: Gravitational-wave parameter estimation with autoregressive neural
network flows
- Authors: Stephen R. Green, Christine Simpson, Jonathan Gair
- Abstract summary: We introduce the use of autoregressive normalizing flows for rapid likelihood-free inference of binary black hole system parameters from gravitational-wave data with deep neural networks.
A normalizing flow is an invertible mapping on a sample space that can be used to induce a transformation from a simple probability distribution to a more complex one.
We build a more powerful latent variable model by incorporating autoregressive flows within the variational autoencoder framework.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce the use of autoregressive normalizing flows for rapid
likelihood-free inference of binary black hole system parameters from
gravitational-wave data with deep neural networks. A normalizing flow is an
invertible mapping on a sample space that can be used to induce a
transformation from a simple probability distribution to a more complex one: if
the simple distribution can be rapidly sampled and its density evaluated, then
so can the complex distribution. Our first application to gravitational waves
uses an autoregressive flow, conditioned on detector strain data, to map a
multivariate standard normal distribution into the posterior distribution over
system parameters. We train the model on artificial strain data consisting of
IMRPhenomPv2 waveforms drawn from a five-parameter $(m_1, m_2, \phi_0, t_c,
d_L)$ prior and stationary Gaussian noise realizations with a fixed power
spectral density. This gives performance comparable to current best
deep-learning approaches to gravitational-wave parameter estimation. We then
build a more powerful latent variable model by incorporating autoregressive
flows within the variational autoencoder framework. This model has performance
comparable to Markov chain Monte Carlo and, in particular, successfully models
the multimodal $\phi_0$ posterior. Finally, we train the autoregressive latent
variable model on an expanded parameter space, including also aligned spins
$(\chi_{1z}, \chi_{2z})$ and binary inclination $\theta_{JN}$, and show that
all parameters and degeneracies are well-recovered. In all cases, sampling is
extremely fast, requiring less than two seconds to draw $10^4$ posterior
samples.
Related papers
- Accelerating Diffusion Models with Parallel Sampling: Inference at Sub-Linear Time Complexity [11.71206628091551]
Diffusion models are costly to train and evaluate, reducing the inference cost for diffusion models remains a major goal.
Inspired by the recent empirical success in accelerating diffusion models via the parallel sampling techniqueciteshih2024parallel, we propose to divide the sampling process into $mathcalO(1)$ blocks with parallelizable Picard iterations within each block.
Our results shed light on the potential of fast and efficient sampling of high-dimensional data on fast-evolving modern large-memory GPU clusters.
arXiv Detail & Related papers (2024-05-24T23:59:41Z) - Learning with Norm Constrained, Over-parameterized, Two-layer Neural Networks [54.177130905659155]
Recent studies show that a reproducing kernel Hilbert space (RKHS) is not a suitable space to model functions by neural networks.
In this paper, we study a suitable function space for over- parameterized two-layer neural networks with bounded norms.
arXiv Detail & Related papers (2024-04-29T15:04:07Z) - Deep adaptive sampling for surrogate modeling without labeled data [4.047684532081032]
We present a deep adaptive sampling method for surrogate modeling ($textDAS2$)
In the parametric setting, the residual loss function can be regarded as an unnormalized probability density function.
New samples match the residual-induced distribution, the refined training set can further reduce the statistical error.
arXiv Detail & Related papers (2024-02-17T13:44:02Z) - Model-adapted Fourier sampling for generative compressed sensing [7.130302992490975]
We study generative compressed sensing when the measurement matrix is randomly subsampled from a unitary matrix.
We construct a model-adapted sampling strategy with an improved sample complexity of $textitO(kd| boldsymbolalpha|_22)$ measurements.
arXiv Detail & Related papers (2023-10-08T03:13:16Z) - Towards Faster Non-Asymptotic Convergence for Diffusion-Based Generative
Models [49.81937966106691]
We develop a suite of non-asymptotic theory towards understanding the data generation process of diffusion models.
In contrast to prior works, our theory is developed based on an elementary yet versatile non-asymptotic approach.
arXiv Detail & Related papers (2023-06-15T16:30:08Z) - Capturing dynamical correlations using implicit neural representations [85.66456606776552]
We develop an artificial intelligence framework which combines a neural network trained to mimic simulated data from a model Hamiltonian with automatic differentiation to recover unknown parameters from experimental data.
In doing so, we illustrate the ability to build and train a differentiable model only once, which then can be applied in real-time to multi-dimensional scattering data.
arXiv Detail & Related papers (2023-04-08T07:55:36Z) - Neural Inference of Gaussian Processes for Time Series Data of Quasars [72.79083473275742]
We introduce a new model that enables it to describe quasar spectra completely.
We also introduce a new method of inference of Gaussian process parameters, which we call $textitNeural Inference$.
The combination of both the CDRW model and Neural Inference significantly outperforms the baseline DRW and MLE.
arXiv Detail & Related papers (2022-11-17T13:01:26Z) - Sampling Approximately Low-Rank Ising Models: MCMC meets Variational
Methods [35.24886589614034]
We consider quadratic definite Ising models on the hypercube with a general interaction $J$.
Our general result implies the first time sampling algorithms for low-rank Ising models.
arXiv Detail & Related papers (2022-02-17T21:43:50Z) - Inverting brain grey matter models with likelihood-free inference: a
tool for trustable cytoarchitecture measurements [62.997667081978825]
characterisation of the brain grey matter cytoarchitecture with quantitative sensitivity to soma density and volume remains an unsolved challenge in dMRI.
We propose a new forward model, specifically a new system of equations, requiring a few relatively sparse b-shells.
We then apply modern tools from Bayesian analysis known as likelihood-free inference (LFI) to invert our proposed model.
arXiv Detail & Related papers (2021-11-15T09:08:27Z) - Bayesian multiscale deep generative model for the solution of
high-dimensional inverse problems [0.0]
A novel multiscale Bayesian inference approach is introduced based on deep probabilistic generative models.
The method allows high-dimensional parameter estimation while exhibiting stability, efficiency and accuracy.
arXiv Detail & Related papers (2021-02-04T11:47:21Z) - Uncertainty Inspired RGB-D Saliency Detection [70.50583438784571]
We propose the first framework to employ uncertainty for RGB-D saliency detection by learning from the data labeling process.
Inspired by the saliency data labeling process, we propose a generative architecture to achieve probabilistic RGB-D saliency detection.
Results on six challenging RGB-D benchmark datasets show our approach's superior performance in learning the distribution of saliency maps.
arXiv Detail & Related papers (2020-09-07T13:01:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.