Flow-based Distributionally Robust Optimization
- URL: http://arxiv.org/abs/2310.19253v4
- Date: Sat, 24 Feb 2024 23:20:28 GMT
- Title: Flow-based Distributionally Robust Optimization
- Authors: Chen Xu, Jonghyeok Lee, Xiuyuan Cheng, Yao Xie
- Abstract summary: We present a framework, called $textttFlowDRO$, for solving flow-based distributionally robust optimization (DRO) problems with Wasserstein uncertainty sets.
We aim to find continuous worst-case distribution (also called the Least Favorable Distribution, LFD) and sample from it.
We demonstrate its usage in adversarial learning, distributionally robust hypothesis testing, and a new mechanism for data-driven distribution perturbation differential privacy.
- Score: 23.232731771848883
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a computationally efficient framework, called $\texttt{FlowDRO}$,
for solving flow-based distributionally robust optimization (DRO) problems with
Wasserstein uncertainty sets while aiming to find continuous worst-case
distribution (also called the Least Favorable Distribution, LFD) and sample
from it. The requirement for LFD to be continuous is so that the algorithm can
be scalable to problems with larger sample sizes and achieve better
generalization capability for the induced robust algorithms. To tackle the
computationally challenging infinitely dimensional optimization problem, we
leverage flow-based models and continuous-time invertible transport maps
between the data distribution and the target distribution and develop a
Wasserstein proximal gradient flow type algorithm. In theory, we establish the
equivalence of the solution by optimal transport map to the original
formulation, as well as the dual form of the problem through Wasserstein
calculus and Brenier theorem. In practice, we parameterize the transport maps
by a sequence of neural networks progressively trained in blocks by gradient
descent. We demonstrate its usage in adversarial learning, distributionally
robust hypothesis testing, and a new mechanism for data-driven distribution
perturbation differential privacy, where the proposed method gives strong
empirical performance on high-dimensional real data.
Related papers
- Straightness of Rectified Flow: A Theoretical Insight into Wasserstein Convergence [54.580605276017096]
Rectified Flow (RF) aims to learn straight flow trajectories from noise to data using a sequence of convex optimization problems.
RF theoretically straightens the trajectory through successive rectifications, reducing the number of evaluations function (NFEs) while sampling.
We provide the first theoretical analysis of the Wasserstein distance between the sampling distribution of RF and the target distribution.
arXiv Detail & Related papers (2024-10-19T02:36:11Z) - Dynamical Measure Transport and Neural PDE Solvers for Sampling [77.38204731939273]
We tackle the task of sampling from a probability density as transporting a tractable density function to the target.
We employ physics-informed neural networks (PINNs) to approximate the respective partial differential equations (PDEs) solutions.
PINNs allow for simulation- and discretization-free optimization and can be trained very efficiently.
arXiv Detail & Related papers (2024-07-10T17:39:50Z) - Distributed Markov Chain Monte Carlo Sampling based on the Alternating
Direction Method of Multipliers [143.6249073384419]
In this paper, we propose a distributed sampling scheme based on the alternating direction method of multipliers.
We provide both theoretical guarantees of our algorithm's convergence and experimental evidence of its superiority to the state-of-the-art.
In simulation, we deploy our algorithm on linear and logistic regression tasks and illustrate its fast convergence compared to existing gradient-based methods.
arXiv Detail & Related papers (2024-01-29T02:08:40Z) - Geometry-Aware Normalizing Wasserstein Flows for Optimal Causal
Inference [0.0]
This paper presents a groundbreaking approach to causal inference by integrating continuous normalizing flows with parametric submodels.
We leverage optimal transport and Wasserstein gradient flows to develop causal inference methodologies with minimal variance in finite-sample settings.
Preliminary experiments showcase our method's superiority, yielding lower mean-squared errors compared to standard flows.
arXiv Detail & Related papers (2023-11-30T18:59:05Z) - Dynamic Flows on Curved Space Generated by Labeled Data [17.621847430986854]
We propose a gradient flow method to generate new samples close to a dataset of interest.
We show that our method can improve the accuracy of classification models in transfer learning settings.
arXiv Detail & Related papers (2023-01-31T19:53:01Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Relative Entropy-Regularized Optimal Transport on a Graph: a new
algorithm and an experimental comparison [0.0]
The present work investigates a new relative entropy-regularized algorithm for solving the optimal transport on a graph problem within the randomized shortest paths formalism.
The main advantage of this new formulation is the fact that it can easily accommodate edge flow capacity constraints.
The resulting optimal routing policy, i.e., the probability distribution of following an edge in each node, is Markovian and is computed by constraining the input and output flows to the prescribed marginal probabilities.
arXiv Detail & Related papers (2021-08-23T08:25:51Z) - Distributed Averaging Methods for Randomized Second Order Optimization [54.51566432934556]
We consider distributed optimization problems where forming the Hessian is computationally challenging and communication is a bottleneck.
We develop unbiased parameter averaging methods for randomized second order optimization that employ sampling and sketching of the Hessian.
We also extend the framework of second order averaging methods to introduce an unbiased distributed optimization framework for heterogeneous computing systems.
arXiv Detail & Related papers (2020-02-16T09:01:18Z) - A Near-Optimal Gradient Flow for Learning Neural Energy-Based Models [93.24030378630175]
We propose a novel numerical scheme to optimize the gradient flows for learning energy-based models (EBMs)
We derive a second-order Wasserstein gradient flow of the global relative entropy from Fokker-Planck equation.
Compared with existing schemes, Wasserstein gradient flow is a smoother and near-optimal numerical scheme to approximate real data densities.
arXiv Detail & Related papers (2019-10-31T02:26:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.