Translation and Rotation Equivariant Normalizing Flow (TRENF) for
Optimal Cosmological Analysis
- URL: http://arxiv.org/abs/2202.05282v1
- Date: Thu, 10 Feb 2022 19:00:03 GMT
- Title: Translation and Rotation Equivariant Normalizing Flow (TRENF) for
Optimal Cosmological Analysis
- Authors: Biwei Dai and Uros Seljak
- Abstract summary: Our universe is homogeneous and isotropic, and its perturbations obey translation and rotation symmetry.
We develop a generative Normalizing Flow model which explicitly incorporates these symmetries.
TRENF gives direct access to the high dimensional data likelihood p(x|y) as a function of the labels y.
- Score: 7.6146285961466
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Our universe is homogeneous and isotropic, and its perturbations obey
translation and rotation symmetry. In this work we develop Translation and
Rotation Equivariant Normalizing Flow (TRENF), a generative Normalizing Flow
(NF) model which explicitly incorporates these symmetries, defining the data
likelihood via a sequence of Fourier space-based convolutions and pixel-wise
nonlinear transforms. TRENF gives direct access to the high dimensional data
likelihood p(x|y) as a function of the labels y, such as cosmological
parameters. In contrast to traditional analyses based on summary statistics,
the NF approach has no loss of information since it preserves the full
dimensionality of the data. On Gaussian random fields, the TRENF likelihood
agrees well with the analytical expression and saturates the Fisher information
content in the labels y. On nonlinear cosmological overdensity fields from
N-body simulations, TRENF leads to significant improvements in constraining
power over the standard power spectrum summary statistic. TRENF is also a
generative model of the data, and we show that TRENF samples agree well with
the N-body simulations it trained on, and that the inverse mapping of the data
agrees well with a Gaussian white noise both visually and on various summary
statistics: when this is perfectly achieved the resulting p(x|y) likelihood
analysis becomes optimal. Finally, we develop a generalization of this model
that can handle effects that break the symmetry of the data, such as the survey
mask, which enables likelihood analysis on data without periodic boundaries.
Related papers
- Unsupervised Representation Learning from Sparse Transformation Analysis [79.94858534887801]
We propose to learn representations from sequence data by factorizing the transformations of the latent variables into sparse components.
Input data are first encoded as distributions of latent activations and subsequently transformed using a probability flow model.
arXiv Detail & Related papers (2024-10-07T23:53:25Z) - Equivariant score-based generative models provably learn distributions with symmetries efficiently [7.90752151686317]
Empirical studies have demonstrated that incorporating symmetries into generative models can provide better generalization and sampling efficiency.
We provide the first theoretical analysis and guarantees of score-based generative models (SGMs) for learning distributions that are invariant with respect to some group symmetry.
arXiv Detail & Related papers (2024-10-02T05:14:28Z) - Fisher Flow Matching for Generative Modeling over Discrete Data [12.69975914345141]
We introduce Fisher-Flow, a novel flow-matching model for discrete data.
Fisher-Flow takes a manifestly geometric perspective by considering categorical distributions over discrete data.
We prove that the gradient flow induced by Fisher-Flow is optimal in reducing the forward KL divergence.
arXiv Detail & Related papers (2024-05-23T15:02:11Z) - Manifold Learning with Sparse Regularised Optimal Transport [0.17205106391379024]
Real-world datasets are subject to noisy observations and sampling, so that distilling information about the underlying manifold is a major challenge.
We propose a method for manifold learning that utilises a symmetric version of optimal transport with a quadratic regularisation.
We prove that the resulting kernel is consistent with a Laplace-type operator in the continuous limit, establish robustness to heteroskedastic noise and exhibit these results in simulations.
arXiv Detail & Related papers (2023-07-19T08:05:46Z) - ManiFlow: Implicitly Representing Manifolds with Normalizing Flows [145.9820993054072]
Normalizing Flows (NFs) are flexible explicit generative models that have been shown to accurately model complex real-world data distributions.
We propose an optimization objective that recovers the most likely point on the manifold given a sample from the perturbed distribution.
Finally, we focus on 3D point clouds for which we utilize the explicit nature of NFs, i.e. surface normals extracted from the gradient of the log-likelihood and the log-likelihood itself.
arXiv Detail & Related papers (2022-08-18T16:07:59Z) - Nonlinear Isometric Manifold Learning for Injective Normalizing Flows [58.720142291102135]
We use isometries to separate manifold learning and density estimation.
We also employ autoencoders to design embeddings with explicit inverses that do not distort the probability distribution.
arXiv Detail & Related papers (2022-03-08T08:57:43Z) - Flow-Based Likelihoods for Non-Gaussian Inference [0.0]
We investigate the use of data-driven likelihoods to bypass a key assumption made in many scientific analyses.
We show that the likelihood can be reconstructed to a precision equal to that of sampling error due to a finite sample size.
By introducing a suite of tests that can capture different levels of NG in the data, we show that the success or failure of traditional data-driven likelihoods can be tied back to the structure of the NG in the data.
arXiv Detail & Related papers (2020-07-10T18:00:00Z) - A Random Matrix Analysis of Random Fourier Features: Beyond the Gaussian
Kernel, a Precise Phase Transition, and the Corresponding Double Descent [85.77233010209368]
This article characterizes the exacts of random Fourier feature (RFF) regression, in the realistic setting where the number of data samples $n$ is all large and comparable.
This analysis also provides accurate estimates of training and test regression errors for large $n,p,N$.
arXiv Detail & Related papers (2020-06-09T02:05:40Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z) - Semi-Supervised Learning with Normalizing Flows [54.376602201489995]
FlowGMM is an end-to-end approach to generative semi supervised learning with normalizing flows.
We show promising results on a wide range of applications, including AG-News and Yahoo Answers text data.
arXiv Detail & Related papers (2019-12-30T17:36:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.