Towards the Semantic Weak Generalization Problem in Generative Zero-Shot
Learning: Ante-hoc and Post-hoc
- URL: http://arxiv.org/abs/2204.11280v1
- Date: Sun, 24 Apr 2022 13:54:42 GMT
- Title: Towards the Semantic Weak Generalization Problem in Generative Zero-Shot
Learning: Ante-hoc and Post-hoc
- Authors: Dubing Chen, Yuming Shen, Haofeng Zhang, Philip H.S. Torr
- Abstract summary: We present a simple and effective strategy lowering the previously unexplored factors that limit the performance ceiling of generative Zero-Shot Learning (ZSL)
We begin by formally defining semantic generalization, then look into approaches for reducing the semantic weak generalization problem.
In the ante-hoc phase, we augment the generator's semantic input, as well as relax the fitting target of the generator.
- Score: 89.68803484284408
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we present a simple and effective strategy lowering the
previously unexplored factors that limit the performance ceiling of generative
Zero-Shot Learning (ZSL). We begin by formally defining semantic
generalization, then look into approaches for reducing the semantic weak
generalization problem and minimizing its negative influence on classifier
training. In the ante-hoc phase, we augment the generator's semantic input, as
well as relax the fitting target of the generator. In the post-hoc phase (after
generating simulated unseen samples), we derive from the gradient of the loss
function to minimize the gradient increment on seen classifier weights carried
by biased unseen distribution, which tends to cause misleading on intra-seen
class decision boundaries. Without complicated designs, our approach hit the
essential problem and significantly outperform the state-of-the-art on four
widely used ZSL datasets.
Related papers
- Semi-Supervised Laplace Learning on Stiefel Manifolds [48.3427853588646]
We develop the framework Sequential Subspace for graph-based, supervised samples at low-label rates.
We achieves that our methods at extremely low rates, and high label rates.
arXiv Detail & Related papers (2023-07-31T20:19:36Z) - Explicit Regularization in Overparametrized Models via Noise Injection [14.492434617004932]
We show that small perturbations induce explicit regularization for simple finite-dimensional models.
We empirically show that the small perturbations lead to better generalization performance than vanilla (stochastic) gradient descent training.
arXiv Detail & Related papers (2022-06-09T17:00:23Z) - Robust Unsupervised Learning via L-Statistic Minimization [38.49191945141759]
We present a general approach to this problem focusing on unsupervised learning.
The key assumption is that the perturbing distribution is characterized by larger losses relative to a given class of admissible models.
We prove uniform convergence bounds with respect to the proposed criterion for several popular models in unsupervised learning.
arXiv Detail & Related papers (2020-12-14T10:36:06Z) - A Simple but Tough-to-Beat Data Augmentation Approach for Natural
Language Understanding and Generation [53.8171136907856]
We introduce a set of simple yet effective data augmentation strategies dubbed cutoff.
cutoff relies on sampling consistency and thus adds little computational overhead.
cutoff consistently outperforms adversarial training and achieves state-of-the-art results on the IWSLT2014 German-English dataset.
arXiv Detail & Related papers (2020-09-29T07:08:35Z) - Generalized Zero-Shot Learning via VAE-Conditioned Generative Flow [83.27681781274406]
Generalized zero-shot learning aims to recognize both seen and unseen classes by transferring knowledge from semantic descriptions to visual representations.
Recent generative methods formulate GZSL as a missing data problem, which mainly adopts GANs or VAEs to generate visual features for unseen classes.
We propose a conditional version of generative flows for GZSL, i.e., VAE-Conditioned Generative Flow (VAE-cFlow)
arXiv Detail & Related papers (2020-09-01T09:12:31Z) - Unbiased Risk Estimators Can Mislead: A Case Study of Learning with
Complementary Labels [92.98756432746482]
We study a weakly supervised problem called learning with complementary labels.
We show that the quality of gradient estimation matters more in risk minimization.
We propose a novel surrogate complementary loss(SCL) framework that trades zero bias with reduced variance.
arXiv Detail & Related papers (2020-07-05T04:19:37Z) - Adversarial Classification via Distributional Robustness with
Wasserstein Ambiguity [12.576828231302134]
Under Wasserstein ambiguity, the model aims to minimize the value-at-risk of misclassification.
We show that, despite the non-marginity of this classification, standard descent methods appear to converger for this problem.
arXiv Detail & Related papers (2020-05-28T07:28:47Z) - Approximation Schemes for ReLU Regression [80.33702497406632]
We consider the fundamental problem of ReLU regression.
The goal is to output the best fitting ReLU with respect to square loss given to draws from some unknown distribution.
arXiv Detail & Related papers (2020-05-26T16:26:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.