How does the Combined Risk Affect the Performance of Unsupervised Domain
Adaptation Approaches?
- URL: http://arxiv.org/abs/2101.01104v1
- Date: Wed, 30 Dec 2020 00:46:57 GMT
- Title: How does the Combined Risk Affect the Performance of Unsupervised Domain
Adaptation Approaches?
- Authors: Li Zhong, Zhen Fang, Feng Liu, Jie Lu, Bo Yuan, Guangquan Zhang
- Abstract summary: Unsupervised domain adaptation (UDA) aims to train a target classifier with labeled samples from the source domain and unlabeled samples from the target domain.
E-MixNet employs enhanced mixup, a generic vicinal distribution, on the labeled source samples and pseudo-labeled target samples to calculate a proxy of the combined risk.
- Score: 33.65954640678556
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unsupervised domain adaptation (UDA) aims to train a target classifier with
labeled samples from the source domain and unlabeled samples from the target
domain. Classical UDA learning bounds show that target risk is upper bounded by
three terms: source risk, distribution discrepancy, and combined risk. Based on
the assumption that the combined risk is a small fixed value, methods based on
this bound train a target classifier by only minimizing estimators of the
source risk and the distribution discrepancy. However, the combined risk may
increase when minimizing both estimators, which makes the target risk
uncontrollable. Hence the target classifier cannot achieve ideal performance if
we fail to control the combined risk. To control the combined risk, the key
challenge takes root in the unavailability of the labeled samples in the target
domain. To address this key challenge, we propose a method named E-MixNet.
E-MixNet employs enhanced mixup, a generic vicinal distribution, on the labeled
source samples and pseudo-labeled target samples to calculate a proxy of the
combined risk. Experiments show that the proxy can effectively curb the
increase of the combined risk when minimizing the source risk and distribution
discrepancy. Furthermore, we show that if the proxy of the combined risk is
added into loss functions of four representative UDA methods, their performance
is also improved.
Related papers
- Data-driven decision-making under uncertainty with entropic risk measure [5.407319151576265]
The entropic risk measure is widely used in high-stakes decision making to account for tail risks associated with an uncertain loss.
To debias the empirical entropic risk estimator, we propose a strongly consistent bootstrapping procedure.
We show that cross validation methods can result in significantly higher out-of-sample risk for the insurer if the bias in validation performance is not corrected for.
arXiv Detail & Related papers (2024-09-30T04:02:52Z) - Data-Adaptive Tradeoffs among Multiple Risks in Distribution-Free Prediction [55.77015419028725]
We develop methods that permit valid control of risk when threshold and tradeoff parameters are chosen adaptively.
Our methodology supports monotone and nearly-monotone risks, but otherwise makes no distributional assumptions.
arXiv Detail & Related papers (2024-03-28T17:28:06Z) - RiskQ: Risk-sensitive Multi-Agent Reinforcement Learning Value Factorization [49.26510528455664]
We introduce the Risk-sensitive Individual-Global-Max (RIGM) principle as a generalization of the Individual-Global-Max (IGM) and Distributional IGM (DIGM) principles.
We show that RiskQ can obtain promising performance through extensive experiments.
arXiv Detail & Related papers (2023-11-03T07:18:36Z) - Domain Generalization without Excess Empirical Risk [83.26052467843725]
A common approach is designing a data-driven surrogate penalty to capture generalization and minimize the empirical risk jointly with the penalty.
We argue that a significant failure mode of this recipe is an excess risk due to an erroneous penalty or hardness in joint optimization.
We present an approach that eliminates this problem. Instead of jointly minimizing empirical risk with the penalty, we minimize the penalty under the constraint of optimality of the empirical risk.
arXiv Detail & Related papers (2023-08-30T08:46:46Z) - A Risk-Sensitive Approach to Policy Optimization [21.684251937825234]
Standard deep reinforcement learning (DRL) aims to maximize expected reward, considering collected experiences equally in formulating a policy.
We propose a more direct approach whereby risk-sensitive objectives, specified in terms of the cumulative distribution function (CDF) of the distribution of full-episode rewards, are optimized.
We demonstrate that the use of moderately "pessimistic" risk profiles, which emphasize scenarios where the agent performs poorly, leads to enhanced exploration and a continual focus on addressing deficiencies.
arXiv Detail & Related papers (2022-08-19T00:55:05Z) - Tailoring to the Tails: Risk Measures for Fine-Grained Tail Sensitivity [10.482805367361818]
Expected risk rearrangement (ERM) is at the core of machine learning systems.
We propose a general approach to construct risk measures which exhibit a desired tail sensitivity.
arXiv Detail & Related papers (2022-08-05T09:51:18Z) - Efficient Risk-Averse Reinforcement Learning [79.61412643761034]
In risk-averse reinforcement learning (RL), the goal is to optimize some risk measure of the returns.
We prove that under certain conditions this inevitably leads to a local-optimum barrier, and propose a soft risk mechanism to bypass it.
We demonstrate improved risk aversion in maze navigation, autonomous driving, and resource allocation benchmarks.
arXiv Detail & Related papers (2022-05-10T19:40:52Z) - Risk Consistent Multi-Class Learning from Label Proportions [64.0125322353281]
This study addresses a multiclass learning from label proportions (MCLLP) setting in which training instances are provided in bags.
Most existing MCLLP methods impose bag-wise constraints on the prediction of instances or assign them pseudo-labels.
A risk-consistent method is proposed for instance classification using the empirical risk minimization framework.
arXiv Detail & Related papers (2022-03-24T03:49:04Z) - SENTINEL: Taming Uncertainty with Ensemble-based Distributional
Reinforcement Learning [6.587644069410234]
We consider risk-sensitive sequential decision-making in model-based reinforcement learning (RL)
We introduce a novel quantification of risk, namely emphcomposite risk
We experimentally verify that SENTINEL-K estimates the return distribution better, and while used with composite risk estimate, demonstrates better risk-sensitive performance than competing RL algorithms.
arXiv Detail & Related papers (2021-02-22T14:45:39Z) - Class-Weighted Classification: Trade-offs and Robust Approaches [38.12414892326325]
We address imbalanced classification, the problem in which a label may have low marginal probability relative to other labels.
We show that particular choices of the weighting set leads to a special instance of conditional value at risk.
We generalize this weighting to derive a new robust risk problem that we call heterogeneous conditional value at risk.
arXiv Detail & Related papers (2020-05-26T16:45:13Z) - Mean-Variance Policy Iteration for Risk-Averse Reinforcement Learning [75.17074235764757]
We present a framework for risk-averse control in a discounted infinite horizon MDP.
MVPI enjoys great flexibility in that any policy evaluation method and risk-neutral control method can be dropped in for risk-averse control off the shelf.
This flexibility reduces the gap between risk-neutral control and risk-averse control and is achieved by working on a novel augmented MDP.
arXiv Detail & Related papers (2020-04-22T22:23:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.