Robust Optimal Transport with Applications in Generative Modeling and
Domain Adaptation
- URL: http://arxiv.org/abs/2010.05862v1
- Date: Mon, 12 Oct 2020 17:13:40 GMT
- Title: Robust Optimal Transport with Applications in Generative Modeling and
Domain Adaptation
- Authors: Yogesh Balaji, Rama Chellappa and Soheil Feizi
- Abstract summary: Optimal Transport (OT) distances such as Wasserstein have been used in several areas such as GANs and domain adaptation.
We propose a computationally-efficient dual form of the robust OT optimization that is amenable to modern deep learning applications.
Our approach can train state-of-the-art GAN models on noisy datasets corrupted with outlier distributions.
- Score: 120.69747175899421
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Optimal Transport (OT) distances such as Wasserstein have been used in
several areas such as GANs and domain adaptation. OT, however, is very
sensitive to outliers (samples with large noise) in the data since in its
objective function, every sample, including outliers, is weighed similarly due
to the marginal constraints. To remedy this issue, robust formulations of OT
with unbalanced marginal constraints have previously been proposed. However,
employing these methods in deep learning problems such as GANs and domain
adaptation is challenging due to the instability of their dual optimization
solvers. In this paper, we resolve these issues by deriving a
computationally-efficient dual form of the robust OT optimization that is
amenable to modern deep learning applications. We demonstrate the effectiveness
of our formulation in two applications of GANs and domain adaptation. Our
approach can train state-of-the-art GAN models on noisy datasets corrupted with
outlier distributions. In particular, our optimization computes weights for
training samples reflecting how difficult it is for those samples to be
generated in the model. In domain adaptation, our robust OT formulation leads
to improved accuracy compared to the standard adversarial adaptation methods.
Our code is available at https://github.com/yogeshbalaji/robustOT.
Related papers
- Unsupervised Domain Adaptation Via Data Pruning [0.0]
We consider the problem from the perspective of unsupervised domain adaptation (UDA)
We propose AdaPrune, a method for UDA whereby training examples are removed to attempt to align the training distribution to that of the target data.
As a method for UDA, we show that AdaPrune outperforms related techniques, and is complementary to other UDA algorithms such as CORAL.
arXiv Detail & Related papers (2024-09-18T15:48:59Z) - Two-Stage ML-Guided Decision Rules for Sequential Decision Making under Uncertainty [55.06411438416805]
Sequential Decision Making under Uncertainty (SDMU) is ubiquitous in many domains such as energy, finance, and supply chains.
Some SDMU are naturally modeled as Multistage Problems (MSPs) but the resulting optimizations are notoriously challenging from a computational standpoint.
This paper introduces a novel approach Two-Stage General Decision Rules (TS-GDR) to generalize the policy space beyond linear functions.
The effectiveness of TS-GDR is demonstrated through an instantiation using Deep Recurrent Neural Networks named Two-Stage Deep Decision Rules (TS-LDR)
arXiv Detail & Related papers (2024-05-23T18:19:47Z) - Enhanced Online Test-time Adaptation with Feature-Weight Cosine Alignment [7.991720491452191]
Online Test-Time Adaptation (OTTA) has emerged as an effective strategy to handle distributional shifts.
This paper introduces a novel cosine alignment optimization approach with a dual-objective loss function.
Our method outperforms state-of-the-art techniques and sets a new benchmark in multiple datasets.
arXiv Detail & Related papers (2024-05-12T05:57:37Z) - Learning Constrained Optimization with Deep Augmented Lagrangian Methods [54.22290715244502]
A machine learning (ML) model is trained to emulate a constrained optimization solver.
This paper proposes an alternative approach, in which the ML model is trained to predict dual solution estimates directly.
It enables an end-to-end training scheme is which the dual objective is as a loss function, and solution estimates toward primal feasibility, emulating a Dual Ascent method.
arXiv Detail & Related papers (2024-03-06T04:43:22Z) - Informative Data Mining for One-Shot Cross-Domain Semantic Segmentation [84.82153655786183]
We propose a novel framework called Informative Data Mining (IDM) to enable efficient one-shot domain adaptation for semantic segmentation.
IDM provides an uncertainty-based selection criterion to identify the most informative samples, which facilitates quick adaptation and reduces redundant training.
Our approach outperforms existing methods and achieves a new state-of-the-art one-shot performance of 56.7%/55.4% on the GTA5/SYNTHIA to Cityscapes adaptation tasks.
arXiv Detail & Related papers (2023-09-25T15:56:01Z) - Sample-Efficient and Surrogate-Based Design Optimization of Underwater Vehicle Hulls [0.4543820534430522]
We show that theBO-LCB algorithm is the most sample-efficient optimization framework and has the best convergence behavior of those considered.
We also show that our DNN-based surrogate model predicts drag force on test data in tight agreement with CFD simulations, with a mean absolute percentage error (MAPE) of 1.85%.
We demonstrate a two-orders-of-magnitude speedup for the design optimization process when the surrogate model is used.
arXiv Detail & Related papers (2023-04-24T19:52:42Z) - Implicit Bayes Adaptation: A Collaborative Transport Approach [25.96406219707398]
We show that domain adaptation is rooted in the intrinsic representations of the respective data, which are inherently lying in a non-linear submanifold embedded in a higher dimensional Euclidean space.
We show that this is tantamount to an implicit Bayesian framework, which we demonstrate to be viable for a more robust and better-performing approach to domain adaptation.
arXiv Detail & Related papers (2023-04-17T14:13:40Z) - Adaptive Self-supervision Algorithms for Physics-informed Neural
Networks [59.822151945132525]
Physics-informed neural networks (PINNs) incorporate physical knowledge from the problem domain as a soft constraint on the loss function.
We study the impact of the location of the collocation points on the trainability of these models.
We propose a novel adaptive collocation scheme which progressively allocates more collocation points to areas where the model is making higher errors.
arXiv Detail & Related papers (2022-07-08T18:17:06Z) - Model Selection for Bayesian Autoencoders [25.619565817793422]
We propose to optimize the distributional sliced-Wasserstein distance between the output of the autoencoder and the empirical data distribution.
We turn our BAE into a generative model by fitting a flexible Dirichlet mixture model in the latent space.
We evaluate our approach qualitatively and quantitatively using a vast experimental campaign on a number of unsupervised learning tasks and show that, in small-data regimes where priors matter, our approach provides state-of-the-art results.
arXiv Detail & Related papers (2021-06-11T08:55:00Z) - Extrapolation for Large-batch Training in Deep Learning [72.61259487233214]
We show that a host of variations can be covered in a unified framework that we propose.
We prove the convergence of this novel scheme and rigorously evaluate its empirical performance on ResNet, LSTM, and Transformer.
arXiv Detail & Related papers (2020-06-10T08:22:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.