A Unified Joint Maximum Mean Discrepancy for Domain Adaptation
- URL: http://arxiv.org/abs/2101.09979v1
- Date: Mon, 25 Jan 2021 09:46:14 GMT
- Title: A Unified Joint Maximum Mean Discrepancy for Domain Adaptation
- Authors: Wei Wang, Baopu Li, Shuhui Yang, Jing Sun, Zhengming Ding, Junyang
Chen, Xiao Dong, Zhihui Wang, Haojie Li
- Abstract summary: This paper theoretically derives a unified form of JMMD that is easy to optimize.
From the revealed unified JMMD, we illustrate that JMMD degrades the feature-label dependence that benefits to classification.
We propose a novel MMD matrix to promote the dependence, and devise a novel label kernel that is robust to label distribution shift.
- Score: 73.44809425486767
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Domain adaptation has received a lot of attention in recent years, and many
algorithms have been proposed with impressive progress. However, it is still
not fully explored concerning the joint probability distribution (P(X, Y))
distance for this problem, since its empirical estimation derived from the
maximum mean discrepancy (joint maximum mean discrepancy, JMMD) will involve
complex tensor-product operator that is hard to manipulate. To solve this
issue, this paper theoretically derives a unified form of JMMD that is easy to
optimize, and proves that the marginal, class conditional and weighted class
conditional probability distribution distances are our special cases with
different label kernels, among which the weighted class conditional one not
only can realize feature alignment across domains in the category level, but
also deal with imbalance dataset using the class prior probabilities. From the
revealed unified JMMD, we illustrate that JMMD degrades the feature-label
dependence (discriminability) that benefits to classification, and it is
sensitive to the label distribution shift when the label kernel is the weighted
class conditional one. Therefore, we leverage Hilbert Schmidt independence
criterion and propose a novel MMD matrix to promote the dependence, and devise
a novel label kernel that is robust to label distribution shift. Finally, we
conduct extensive experiments on several cross-domain datasets to demonstrate
the validity and effectiveness of the revealed theoretical results.
Related papers
- SimPro: A Simple Probabilistic Framework Towards Realistic Long-Tailed Semi-Supervised Learning [49.94607673097326]
We propose a highly adaptable framework, designated as SimPro, which does not rely on any predefined assumptions about the distribution of unlabeled data.
Our framework, grounded in a probabilistic model, innovatively refines the expectation-maximization algorithm.
Our method showcases consistent state-of-the-art performance across diverse benchmarks and data distribution scenarios.
arXiv Detail & Related papers (2024-02-21T03:39:04Z) - Anomaly Detection Under Uncertainty Using Distributionally Robust
Optimization Approach [0.9217021281095907]
Anomaly detection is defined as the problem of finding data points that do not follow the patterns of the majority.
The one-class Support Vector Machines (SVM) method aims to find a decision boundary to distinguish between normal data points and anomalies.
A distributionally robust chance-constrained model is proposed in which the probability of misclassification is low.
arXiv Detail & Related papers (2023-12-03T06:13:22Z) - Out-Of-Domain Unlabeled Data Improves Generalization [0.7589678255312519]
We propose a novel framework for incorporating unlabeled data into semi-supervised classification problems.
We show that unlabeled samples can be harnessed to narrow the generalization gap.
We validate our claims through experiments conducted on a variety of synthetic and real-world datasets.
arXiv Detail & Related papers (2023-09-29T02:00:03Z) - Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - Class-Distribution-Aware Pseudo Labeling for Semi-Supervised Multi-Label
Learning [97.88458953075205]
Pseudo-labeling has emerged as a popular and effective approach for utilizing unlabeled data.
This paper proposes a novel solution called Class-Aware Pseudo-Labeling (CAP) that performs pseudo-labeling in a class-aware manner.
arXiv Detail & Related papers (2023-05-04T12:52:18Z) - MaxMatch: Semi-Supervised Learning with Worst-Case Consistency [149.03760479533855]
We propose a worst-case consistency regularization technique for semi-supervised learning (SSL)
We present a generalization bound for SSL consisting of the empirical loss terms observed on labeled and unlabeled training data separately.
Motivated by this bound, we derive an SSL objective that minimizes the largest inconsistency between an original unlabeled sample and its multiple augmented variants.
arXiv Detail & Related papers (2022-09-26T12:04:49Z) - From Optimality to Robustness: Dirichlet Sampling Strategies in
Stochastic Bandits [0.0]
We study a generic Dirichlet Sampling (DS) algorithm, based on pairwise comparisons of empirical indices computed with re-sampling of the arms' observations.
We show that different variants of this strategy achieve provably optimal regret guarantees when the distributions are bounded and logarithmic regret for semi-bounded distributions with a mild quantile condition.
arXiv Detail & Related papers (2021-11-18T14:34:21Z) - Conditional Bures Metric for Domain Adaptation [14.528711361447712]
Unsupervised domain adaptation (UDA) has attracted widespread attention in recent years.
Previous UDA methods assume the marginal distributions of different domains are shifted while ignoring the discriminant information in the label distributions.
In this work, we focus on the conditional distribution shift problem which is of great concern to current conditional invariant models.
arXiv Detail & Related papers (2021-07-31T18:06:31Z) - Towards Discriminability and Diversity: Batch Nuclear-norm Maximization
under Label Insufficient Situations [154.51144248210338]
Batch Nuclear-norm Maximization (BNM) is proposed to boost the learning under label insufficient learning scenarios.
BNM outperforms competitors and works well with existing well-known methods.
arXiv Detail & Related papers (2020-03-27T05:04:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.