Unifying Distributionally Robust Optimization via Optimal Transport
Theory
- URL: http://arxiv.org/abs/2308.05414v1
- Date: Thu, 10 Aug 2023 08:17:55 GMT
- Title: Unifying Distributionally Robust Optimization via Optimal Transport
Theory
- Authors: Jose Blanchet, Daniel Kuhn, Jiajin Li, Bahar Taskesen
- Abstract summary: This paper introduces a novel approach that unifies these methods into a single framework based on optimal transport.
Our proposed approach makes it possible for optimal adversarial distributions to simultaneously perturb likelihood and outcomes.
The paper investigates several duality results and presents tractable reformulations that enhance the practical applicability of this unified framework.
- Score: 13.19058156672392
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the past few years, there has been considerable interest in two prominent
approaches for Distributionally Robust Optimization (DRO): Divergence-based and
Wasserstein-based methods. The divergence approach models misspecification in
terms of likelihood ratios, while the latter models it through a measure of
distance or cost in actual outcomes. Building upon these advances, this paper
introduces a novel approach that unifies these methods into a single framework
based on optimal transport (OT) with conditional moment constraints. Our
proposed approach, for example, makes it possible for optimal adversarial
distributions to simultaneously perturb likelihood and outcomes, while
producing an optimal (in an optimal transport sense) coupling between the
baseline model and the adversarial model.Additionally, the paper investigates
several duality results and presents tractable reformulations that enhance the
practical applicability of this unified framework.
Related papers
- Distributionally and Adversarially Robust Logistic Regression via Intersecting Wasserstein Balls [8.720733751119994]
Adversarially robust optimization (ARO) has become the de facto standard for training models to defend against adversarial attacks during testing.
Despite their robustness, these models often suffer from severe overfitting.
We propose two approaches to replace the empirical distribution in training with: (i) a worst-case distribution within an ambiguity set; or (ii) a mixture of the empirical distribution with one derived from an auxiliary dataset.
arXiv Detail & Related papers (2024-07-18T15:59:37Z) - Constrained Bayesian Optimization Under Partial Observations: Balanced
Improvements and Provable Convergence [6.461785985849886]
We endeavor to design an efficient and provable method for expensive POCOPs under the framework of constrained Bayesian optimization.
We present an improved design of the acquisition functions that introduces balanced exploration during optimization.
We propose a Gaussian process embedding different likelihoods as the surrogate model for a partially observable constraint.
arXiv Detail & Related papers (2023-12-06T01:00:07Z) - Improving Transferability of Adversarial Examples via Bayesian Attacks [84.90830931076901]
We introduce a novel extension by incorporating the Bayesian formulation into the model input as well, enabling the joint diversification of both the model input and model parameters.
Our method achieves a new state-of-the-art on transfer-based attacks, improving the average success rate on ImageNet and CIFAR-10 by 19.14% and 2.08%, respectively.
arXiv Detail & Related papers (2023-07-21T03:43:07Z) - Variational Refinement for Importance Sampling Using the Forward
Kullback-Leibler Divergence [77.06203118175335]
Variational Inference (VI) is a popular alternative to exact sampling in Bayesian inference.
Importance sampling (IS) is often used to fine-tune and de-bias the estimates of approximate Bayesian inference procedures.
We propose a novel combination of optimization and sampling techniques for approximate Bayesian inference.
arXiv Detail & Related papers (2021-06-30T11:00:24Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z) - Comparing Probability Distributions with Conditional Transport [63.11403041984197]
We propose conditional transport (CT) as a new divergence and approximate it with the amortized CT (ACT) cost.
ACT amortizes the computation of its conditional transport plans and comes with unbiased sample gradients that are straightforward to compute.
On a wide variety of benchmark datasets generative modeling, substituting the default statistical distance of an existing generative adversarial network with ACT is shown to consistently improve the performance.
arXiv Detail & Related papers (2020-12-28T05:14:22Z) - Deep Ordinal Regression using Optimal Transport Loss and Unimodal Output
Probabilities [3.169089186688223]
We propose a framework for deep ordinal regression based on unimodal output distribution and optimal transport loss.
We empirically analyze the different components of our proposed approach and demonstrate their contribution to the performance of the model.
Experimental results on eight real-world datasets demonstrate that our proposed approach consistently performs on par with and often better than several recently proposed deep learning approaches.
arXiv Detail & Related papers (2020-11-15T19:19:40Z) - Robust, Accurate Stochastic Optimization for Variational Inference [68.83746081733464]
We show that common optimization methods lead to poor variational approximations if the problem is moderately large.
Motivated by these findings, we develop a more robust and accurate optimization framework by viewing the underlying algorithm as producing a Markov chain.
arXiv Detail & Related papers (2020-09-01T19:12:11Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.