Rethink Maximum Mean Discrepancy for Domain Adaptation
- URL: http://arxiv.org/abs/2007.00689v1
- Date: Wed, 1 Jul 2020 18:25:10 GMT
- Title: Rethink Maximum Mean Discrepancy for Domain Adaptation
- Authors: Wei Wang and Haojie Li and Zhengming Ding and Zhihui Wang
- Abstract summary: This paper theoretically proves two essential facts: 1) minimizing the Maximum Mean Discrepancy equals to maximize the source and target intra-class distances respectively but jointly minimize their variance with some implicit weights, so that the feature discriminability degrades.
Experiments on several benchmark datasets not only prove the validity of theoretical results but also demonstrate that our approach could perform better than the comparative state-of-art methods substantially.
- Score: 77.2560592127872
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing domain adaptation methods aim to reduce the distributional
difference between the source and target domains and respect their specific
discriminative information, by establishing the Maximum Mean Discrepancy (MMD)
and the discriminative distances. However, they usually accumulate to consider
those statistics and deal with their relationships by estimating parameters
blindly. This paper theoretically proves two essential facts: 1) minimizing the
MMD equals to maximize the source and target intra-class distances respectively
but jointly minimize their variance with some implicit weights, so that the
feature discriminability degrades; 2) the relationship between the intra-class
and inter-class distances is as one falls, another rises. Based on this, we
propose a novel discriminative MMD. On one hand, we consider the intra-class
and inter-class distances alone to remove a redundant parameter, and the
revealed weights provide their approximate optimal ranges. On the other hand,
we design two different strategies to boost the feature discriminability: 1) we
directly impose a trade-off parameter on the implicit intra-class distance in
MMD to regulate its change; 2) we impose the similar weights revealed in MMD on
inter-class distance and maximize it, then a balanced factor could be
introduced to quantitatively leverage the relative importance between the
feature transferability and its discriminability. The experiments on several
benchmark datasets not only prove the validity of theoretical results but also
demonstrate that our approach could perform better than the comparative
state-of-art methods substantially.
Related papers
- Does Machine Bring in Extra Bias in Learning? Approximating Fairness in Models Promptly [2.002741592555996]
Existing techniques for assessing the discrimination level of machine learning models include commonly used group and individual fairness measures.
We propose a "harmonic fairness measure via manifold (HFM)" based on distances between sets.
Empirical results indicate that the proposed fairness measure HFM is valid and that the proposed ApproxDist is effective and efficient.
arXiv Detail & Related papers (2024-05-15T11:07:40Z) - Domain-Specific Risk Minimization for Out-of-Distribution Generalization [104.17683265084757]
We first establish a generalization bound that explicitly considers the adaptivity gap.
We propose effective gap estimation methods for guiding the selection of a better hypothesis for the target.
The other method is minimizing the gap directly by adapting model parameters using online target samples.
arXiv Detail & Related papers (2022-08-18T06:42:49Z) - Cycle Consistent Probability Divergences Across Different Spaces [38.43511529063335]
Discrepancy measures between probability distributions are at the core of statistical inference and machine learning.
This work proposes a novel unbalanced Monge optimal transport formulation for matching, up to isometries, distributions on different spaces.
arXiv Detail & Related papers (2021-11-22T16:35:58Z) - Keep it Tighter -- A Story on Analytical Mean Embeddings [0.6445605125467574]
Kernel techniques are among the most popular and flexible approaches in data science.
Mean embedding gives rise to a divergence measure referred to as maximum mean discrepancy (MMD)
In this paper we focus on the problem of MMD estimation when the mean embedding of one of the underlying distributions is available analytically.
arXiv Detail & Related papers (2021-10-15T21:29:27Z) - A Unified Joint Maximum Mean Discrepancy for Domain Adaptation [73.44809425486767]
This paper theoretically derives a unified form of JMMD that is easy to optimize.
From the revealed unified JMMD, we illustrate that JMMD degrades the feature-label dependence that benefits to classification.
We propose a novel MMD matrix to promote the dependence, and devise a novel label kernel that is robust to label distribution shift.
arXiv Detail & Related papers (2021-01-25T09:46:14Z) - Fundamental Limits and Tradeoffs in Invariant Representation Learning [99.2368462915979]
Many machine learning applications involve learning representations that achieve two competing goals.
Minimax game-theoretic formulation represents a fundamental tradeoff between accuracy and invariance.
We provide an information-theoretic analysis of this general and important problem under both classification and regression settings.
arXiv Detail & Related papers (2020-12-19T15:24:04Z) - Towards Certified Robustness of Distance Metric Learning [53.96113074344632]
We advocate imposing an adversarial margin in the input space so as to improve the generalization and robustness of metric learning algorithms.
We show that the enlarged margin is beneficial to the generalization ability by using the theoretical technique of algorithmic robustness.
arXiv Detail & Related papers (2020-06-10T16:51:53Z) - Neural Methods for Point-wise Dependency Estimation [129.93860669802046]
We focus on estimating point-wise dependency (PD), which quantitatively measures how likely two outcomes co-occur.
We demonstrate the effectiveness of our approaches in 1) MI estimation, 2) self-supervised representation learning, and 3) cross-modal retrieval task.
arXiv Detail & Related papers (2020-06-09T23:26:15Z) - Maximum Density Divergence for Domain Adaptation [0.0]
Unsupervised domain adaptation addresses the problem of transferring knowledge from a well-labeled source domain to an unlabeled target domain.
We propose a new domain adaptation method named Adversarial Tight Match (ATM) which enjoys the benefits of both adversarial training and metric learning.
arXiv Detail & Related papers (2020-04-27T07:35:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.