Non-generative Generalized Zero-shot Learning via Task-correlated
Disentanglement and Controllable Samples Synthesis
- URL: http://arxiv.org/abs/2203.05335v2
- Date: Sun, 13 Mar 2022 08:19:22 GMT
- Title: Non-generative Generalized Zero-shot Learning via Task-correlated
Disentanglement and Controllable Samples Synthesis
- Authors: Yaogong Feng, Xiaowen Huang, Pengbo Yang, Jian Yu, Jitao Sang
- Abstract summary: We propose a non-generative model to address these problems.
In addition, we formulate a new ZSL task named the 'Few-shot Seen class and Zero-shot Unseen class learning' (FSZU)
- Score: 20.34562156468408
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Synthesizing pseudo samples is currently the most effective way to solve the
Generalized Zero Shot Learning (GZSL) problem. Most models achieve competitive
performance but still suffer from two problems: (1) Feature confounding, the
overall representations confound task-correlated and task-independent features,
and existing models disentangle them in a generative way, but they are
unreasonable to synthesize reliable pseudo samples with limited samples; (2)
Distribution uncertainty, that massive data is needed when existing models
synthesize samples from the uncertain distribution, which causes poor
performance in limited samples of seen classes. In this paper, we propose a
non-generative model to address these problems correspondingly in two modules:
(1) Task-correlated feature disentanglement, to exclude the task-correlated
features from task-independent ones by adversarial learning of domain adaption
towards reasonable synthesis; (2) Controllable pseudo sample synthesis, to
synthesize edge-pseudo and center-pseudo samples with certain characteristics
towards more diversity generated and intuitive transfer. In addation, to
describe the new scene that is the limit seen class samples in the training
process, we further formulate a new ZSL task named the 'Few-shot Seen class and
Zero-shot Unseen class learning' (FSZU). Extensive experiments on four
benchmarks verify that the proposed method is competitive in the GZSL and the
FSZU tasks.
Related papers
- Foster Adaptivity and Balance in Learning with Noisy Labels [26.309508654960354]
We propose a novel approach named textbfSED to deal with label noise in a textbfSelf-adaptivtextbfE and class-balancetextbfD manner.
A mean-teacher model is then employed to correct labels of noisy samples.
We additionally propose a self-adaptive and class-balanced sample re-weighting mechanism to assign different weights to detected noisy samples.
arXiv Detail & Related papers (2024-07-03T03:10:24Z) - Learning with Noisy Labels Using Collaborative Sample Selection and
Contrastive Semi-Supervised Learning [76.00798972439004]
Collaborative Sample Selection (CSS) removes noisy samples from identified clean set.
We introduce a co-training mechanism with a contrastive loss in semi-supervised learning.
arXiv Detail & Related papers (2023-10-24T05:37:20Z) - Hard Sample Matters a Lot in Zero-Shot Quantization [52.32914196337281]
Zero-shot quantization (ZSQ) is promising for compressing and accelerating deep neural networks when the data for training full-precision models are inaccessible.
In ZSQ, network quantization is performed using synthetic samples, thus, the performance of quantized models depends heavily on the quality of synthetic samples.
We propose HArd sample Synthesizing and Training (HAST) to address this issue.
arXiv Detail & Related papers (2023-03-24T06:22:57Z) - Hardness of Samples Need to be Quantified for a Reliable Evaluation
System: Exploring Potential Opportunities with a New Task [24.6240575061124]
Evaluation of models on benchmarks is unreliable without knowing the degree of sample hardness.
We propose a Data Scoring task that requires assignment of each unannotated sample in a benchmark a score between 0 to 1.
arXiv Detail & Related papers (2022-10-14T08:26:32Z) - Centrality and Consistency: Two-Stage Clean Samples Identification for
Learning with Instance-Dependent Noisy Labels [87.48541631675889]
We propose a two-stage clean samples identification method.
First, we employ a class-level feature clustering procedure for the early identification of clean samples.
Second, for the remaining clean samples that are close to the ground truth class boundary, we propose a novel consistency-based classification method.
arXiv Detail & Related papers (2022-07-29T04:54:57Z) - Approximating Constraint Manifolds Using Generative Models for
Sampling-Based Constrained Motion Planning [8.924344714683814]
This paper presents a learning-based sampling strategy for constrained motion planning problems.
We use Conditional Variversaational Autoencoder (CVAE) and Conditional Generative Adrial Net (CGAN) to generate constraint-satisfying sample configurations.
We evaluate the efficiency of these two generative models in terms of their sampling accuracy and coverage of sampling distribution.
arXiv Detail & Related papers (2022-04-14T07:08:30Z) - Generalization of Neural Combinatorial Solvers Through the Lens of
Adversarial Robustness [68.97830259849086]
Most datasets only capture a simpler subproblem and likely suffer from spurious features.
We study adversarial robustness - a local generalization property - to reveal hard, model-specific instances and spurious features.
Unlike in other applications, where perturbation models are designed around subjective notions of imperceptibility, our perturbation models are efficient and sound.
Surprisingly, with such perturbations, a sufficiently expressive neural solver does not suffer from the limitations of the accuracy-robustness trade-off common in supervised learning.
arXiv Detail & Related papers (2021-10-21T07:28:11Z) - Jo-SRC: A Contrastive Approach for Combating Noisy Labels [58.867237220886885]
We propose a noise-robust approach named Jo-SRC (Joint Sample Selection and Model Regularization based on Consistency)
Specifically, we train the network in a contrastive learning manner. Predictions from two different views of each sample are used to estimate its "likelihood" of being clean or out-of-distribution.
arXiv Detail & Related papers (2021-03-24T07:26:07Z) - Lifelong Learning Without a Task Oracle [13.331659934508764]
Supervised deep neural networks are known to undergo a sharp decline in the accuracy of older tasks when new tasks are learned.
We propose and compare several candidate task-assigning mappers which require very little memory overhead.
Best-performing variants only impose an average cost of 1.7% parameter memory increase.
arXiv Detail & Related papers (2020-11-09T21:30:31Z) - Robust Finite Mixture Regression for Heterogeneous Targets [70.19798470463378]
We propose an FMR model that finds sample clusters and jointly models multiple incomplete mixed-type targets simultaneously.
We provide non-asymptotic oracle performance bounds for our model under a high-dimensional learning framework.
The results show that our model can achieve state-of-the-art performance.
arXiv Detail & Related papers (2020-10-12T03:27:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.