Deep importance sampling using tensor trains with application to a
priori and a posteriori rare event estimation
- URL: http://arxiv.org/abs/2209.01941v2
- Date: Thu, 25 May 2023 01:03:33 GMT
- Title: Deep importance sampling using tensor trains with application to a
priori and a posteriori rare event estimation
- Authors: Tiangang Cui, Sergey Dolgov, Robert Scheichl
- Abstract summary: We propose a deep importance sampling method that is suitable for estimating rare event probabilities in high-dimensional problems.
We approximate the optimal importance distribution in a general importance sampling problem as the pushforward of a reference distribution under a composition of order-preserving transformations.
The squared tensor-train decomposition provides a scalable ansatz for building order-preserving high-dimensional transformations via density approximations.
- Score: 2.4815579733050153
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a deep importance sampling method that is suitable for estimating
rare event probabilities in high-dimensional problems. We approximate the
optimal importance distribution in a general importance sampling problem as the
pushforward of a reference distribution under a composition of order-preserving
transformations, in which each transformation is formed by a squared
tensor-train decomposition. The squared tensor-train decomposition provides a
scalable ansatz for building order-preserving high-dimensional transformations
via density approximations. The use of composition of maps moving along a
sequence of bridging densities alleviates the difficulty of directly
approximating concentrated density functions. To compute expectations over
unnormalized probability distributions, we design a ratio estimator that
estimates the normalizing constant using a separate importance distribution,
again constructed via a composition of transformations in tensor-train format.
This offers better theoretical variance reduction compared with self-normalized
importance sampling, and thus opens the door to efficient computation of rare
event probabilities in Bayesian inference problems. Numerical experiments on
problems constrained by differential equations show little to no increase in
the computational complexity with the event probability going to zero, and
allow to compute hitherto unattainable estimates of rare event probabilities
for complex, high-dimensional posterior densities.
Related papers
- Unveiling the Statistical Foundations of Chain-of-Thought Prompting Methods [59.779795063072655]
Chain-of-Thought (CoT) prompting and its variants have gained popularity as effective methods for solving multi-step reasoning problems.
We analyze CoT prompting from a statistical estimation perspective, providing a comprehensive characterization of its sample complexity.
arXiv Detail & Related papers (2024-08-25T04:07:18Z) - TERM Model: Tensor Ring Mixture Model for Density Estimation [48.622060998018206]
In this paper, we take tensor ring decomposition for density estimator, which significantly reduces the number of permutation candidates.
A mixture model that incorporates multiple permutation candidates with adaptive weights is further designed, resulting in increased expressive flexibility.
This approach acknowledges that suboptimal permutations can offer distinctive information besides that of optimal permutations.
arXiv Detail & Related papers (2023-12-13T11:39:56Z) - Instance-Dependent Generalization Bounds via Optimal Transport [51.71650746285469]
Existing generalization bounds fail to explain crucial factors that drive the generalization of modern neural networks.
We derive instance-dependent generalization bounds that depend on the local Lipschitz regularity of the learned prediction function in the data space.
We empirically analyze our generalization bounds for neural networks, showing that the bound values are meaningful and capture the effect of popular regularization methods during training.
arXiv Detail & Related papers (2022-11-02T16:39:42Z) - Nonparametric Probabilistic Regression with Coarse Learners [1.8275108630751844]
We show that we can compute precise conditional densities with minimal assumptions on the shape or form of the density.
We demonstrate this approach on a variety of datasets and show competitive performance, particularly on larger datasets.
arXiv Detail & Related papers (2022-10-28T16:25:26Z) - Reliable amortized variational inference with physics-based latent
distribution correction [0.4588028371034407]
A neural network is trained to approximate the posterior distribution over existing pairs of model and data.
The accuracy of this approach relies on the availability of high-fidelity training data.
We show that our correction step improves the robustness of amortized variational inference with respect to changes in number of source experiments, noise variance, and shifts in the prior distribution.
arXiv Detail & Related papers (2022-07-24T02:38:54Z) - Error Analysis of Tensor-Train Cross Approximation [88.83467216606778]
We provide accuracy guarantees in terms of the entire tensor for both exact and noisy measurements.
Results are verified by numerical experiments, and may have important implications for the usefulness of cross approximations for high-order tensors.
arXiv Detail & Related papers (2022-07-09T19:33:59Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Alternating linear scheme in a Bayesian framework for low-rank tensor
approximation [5.833272638548154]
We find a low-rank representation for a given tensor by solving a Bayesian inference problem.
We present an algorithm that performs the unscented transform in tensor train format.
arXiv Detail & Related papers (2020-12-21T10:15:30Z) - Deep composition of tensor-trains using squared inverse Rosenblatt
transports [0.6091702876917279]
This paper generalises the functional tensor-train approximation of the inverse Rosenblatt transport.
We develop an efficient procedure to compute this transport from a squared tensor-train decomposition.
The resulting deep inverse Rosenblatt transport significantly expands the capability of tensor approximations and transport maps to random variables.
arXiv Detail & Related papers (2020-07-14T11:04:18Z) - Optimal Change-Point Detection with Training Sequences in the Large and
Moderate Deviations Regimes [72.68201611113673]
This paper investigates a novel offline change-point detection problem from an information-theoretic perspective.
We assume that the knowledge of the underlying pre- and post-change distributions are not known and can only be learned from the training sequences which are available.
arXiv Detail & Related papers (2020-03-13T23:39:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.