Debias Coarsely, Sample Conditionally: Statistical Downscaling through
Optimal Transport and Probabilistic Diffusion Models
- URL: http://arxiv.org/abs/2305.15618v2
- Date: Mon, 30 Oct 2023 21:41:14 GMT
- Title: Debias Coarsely, Sample Conditionally: Statistical Downscaling through
Optimal Transport and Probabilistic Diffusion Models
- Authors: Zhong Yi Wan, Ricardo Baptista, Yi-fan Chen, John Anderson, Anudhyan
Boral, Fei Sha, Leonardo Zepeda-N\'u\~nez
- Abstract summary: We introduce a two-stage probabilistic framework for statistical downscaling using unpaired data.
We demonstrate the utility of the proposed approach on one- and two-dimensional fluid flow problems.
- Score: 15.623456909553786
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce a two-stage probabilistic framework for statistical downscaling
using unpaired data. Statistical downscaling seeks a probabilistic map to
transform low-resolution data from a biased coarse-grained numerical scheme to
high-resolution data that is consistent with a high-fidelity scheme. Our
framework tackles the problem by composing two transformations: (i) a debiasing
step via an optimal transport map, and (ii) an upsampling step achieved by a
probabilistic diffusion model with a posteriori conditional sampling. This
approach characterizes a conditional distribution without needing paired data,
and faithfully recovers relevant physical statistics from biased samples. We
demonstrate the utility of the proposed approach on one- and two-dimensional
fluid flow problems, which are representative of the core difficulties present
in numerical simulations of weather and climate. Our method produces realistic
high-resolution outputs from low-resolution inputs, by upsampling resolutions
of 8x and 16x. Moreover, our procedure correctly matches the statistics of
physical quantities, even when the low-frequency content of the inputs and
outputs do not match, a crucial but difficult-to-satisfy assumption needed by
current state-of-the-art alternatives. Code for this work is available at:
https://github.com/google-research/swirl-dynamics/tree/main/swirl_dynamics/projects/probabilistic_di ffusion.
Related papers
- Probabilistic Iterative Hard Thresholding for Sparse Learning [2.5782973781085383]
We present an approach towards solving expectation objective optimization problems with cardinality constraints.
We prove convergence of the underlying process, and demonstrate the performance on two Machine Learning problems.
arXiv Detail & Related papers (2024-09-02T18:14:45Z) - Dynamical Measure Transport and Neural PDE Solvers for Sampling [77.38204731939273]
We tackle the task of sampling from a probability density as transporting a tractable density function to the target.
We employ physics-informed neural networks (PINNs) to approximate the respective partial differential equations (PDEs) solutions.
PINNs allow for simulation- and discretization-free optimization and can be trained very efficiently.
arXiv Detail & Related papers (2024-07-10T17:39:50Z) - On the Trajectory Regularity of ODE-based Diffusion Sampling [79.17334230868693]
Diffusion-based generative models use differential equations to establish a smooth connection between a complex data distribution and a tractable prior distribution.
In this paper, we identify several intriguing trajectory properties in the ODE-based sampling process of diffusion models.
arXiv Detail & Related papers (2024-05-18T15:59:41Z) - Flow-based Distributionally Robust Optimization [23.232731771848883]
We present a framework, called $textttFlowDRO$, for solving flow-based distributionally robust optimization (DRO) problems with Wasserstein uncertainty sets.
We aim to find continuous worst-case distribution (also called the Least Favorable Distribution, LFD) and sample from it.
We demonstrate its usage in adversarial learning, distributionally robust hypothesis testing, and a new mechanism for data-driven distribution perturbation differential privacy.
arXiv Detail & Related papers (2023-10-30T03:53:31Z) - Arbitrary Distributions Mapping via SyMOT-Flow: A Flow-based Approach Integrating Maximum Mean Discrepancy and Optimal Transport [2.7309692684728617]
We introduce a novel model called SyMOT-Flow that trains an invertible transformation by minimizing the symmetric maximum mean discrepancy between samples from two unknown distributions.
The resulting transformation leads to more stable and accurate sample generation.
arXiv Detail & Related papers (2023-08-26T08:39:16Z) - User-defined Event Sampling and Uncertainty Quantification in Diffusion
Models for Physical Dynamical Systems [49.75149094527068]
We show that diffusion models can be adapted to make predictions and provide uncertainty quantification for chaotic dynamical systems.
We develop a probabilistic approximation scheme for the conditional score function which converges to the true distribution as the noise level decreases.
We are able to sample conditionally on nonlinear userdefined events at inference time, and matches data statistics even when sampling from the tails of the distribution.
arXiv Detail & Related papers (2023-06-13T03:42:03Z) - Learning Unnormalized Statistical Models via Compositional Optimization [73.30514599338407]
Noise-contrastive estimation(NCE) has been proposed by formulating the objective as the logistic loss of the real data and the artificial noise.
In this paper, we study it a direct approach for optimizing the negative log-likelihood of unnormalized models.
arXiv Detail & Related papers (2023-06-13T01:18:16Z) - Sampling from Arbitrary Functions via PSD Models [55.41644538483948]
We take a two-step approach by first modeling the probability distribution and then sampling from that model.
We show that these models can approximate a large class of densities concisely using few evaluations, and present a simple algorithm to effectively sample from these models.
arXiv Detail & Related papers (2021-10-20T12:25:22Z) - Low-rank statistical finite elements for scalable model-data synthesis [0.8602553195689513]
statFEM acknowledges a priori model misspecification, by embedding forcing within the governing equations.
The method reconstructs the observed data-generating processes with minimal loss of information.
This article overcomes this hurdle by embedding a low-rank approximation of the underlying dense covariance matrix.
arXiv Detail & Related papers (2021-09-10T09:51:43Z) - Comparing Probability Distributions with Conditional Transport [63.11403041984197]
We propose conditional transport (CT) as a new divergence and approximate it with the amortized CT (ACT) cost.
ACT amortizes the computation of its conditional transport plans and comes with unbiased sample gradients that are straightforward to compute.
On a wide variety of benchmark datasets generative modeling, substituting the default statistical distance of an existing generative adversarial network with ACT is shown to consistently improve the performance.
arXiv Detail & Related papers (2020-12-28T05:14:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.