Bi-directional Distribution Alignment for Transductive Zero-Shot
Learning
- URL: http://arxiv.org/abs/2303.08698v2
- Date: Sun, 19 Mar 2023 08:00:20 GMT
- Title: Bi-directional Distribution Alignment for Transductive Zero-Shot
Learning
- Authors: Zhicai Wang, Yanbin Hao, Tingting Mu, Ouxiang Li, Shuo Wang, Xiangnan
He
- Abstract summary: We propose a novel zero-shot learning model (TZSL) called Bi-VAEGAN.
It largely improves the shift by a strengthened distribution alignment between the visual and auxiliary spaces.
In benchmark evaluation, Bi-VAEGAN achieves the new state of the arts under both the standard and generalized TZSL settings.
- Score: 48.80413182126543
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is well-known that zero-shot learning (ZSL) can suffer severely from the
problem of domain shift, where the true and learned data distributions for the
unseen classes do not match. Although transductive ZSL (TZSL) attempts to
improve this by allowing the use of unlabelled examples from the unseen
classes, there is still a high level of distribution shift. We propose a novel
TZSL model (named as Bi-VAEGAN), which largely improves the shift by a
strengthened distribution alignment between the visual and auxiliary spaces.
The key proposal of the model design includes (1) a bi-directional distribution
alignment, (2) a simple but effective L_2-norm based feature normalization
approach, and (3) a more sophisticated unseen class prior estimation approach.
In benchmark evaluation using four datasets, Bi-VAEGAN achieves the new state
of the arts under both the standard and generalized TZSL settings. Code could
be found at https://github.com/Zhicaiwww/Bi-VAEGAN
Related papers
- An Iterative Co-Training Transductive Framework for Zero Shot Learning [24.401200814880124]
We introduce an iterative co-training framework which contains two different base ZSL models and an exchanging module.
At each iteration, the two different ZSL models are co-trained to separately predict pseudo labels for the unseen-class samples.
Our framework can gradually boost the ZSL performance by fully exploiting the potential complementarity of the two models' classification capabilities.
arXiv Detail & Related papers (2022-03-30T04:08:44Z) - A Strong Baseline for Semi-Supervised Incremental Few-Shot Learning [54.617688468341704]
Few-shot learning aims to learn models that generalize to novel classes with limited training samples.
We propose a novel paradigm containing two parts: (1) a well-designed meta-training algorithm for mitigating ambiguity between base and novel classes caused by unreliable pseudo labels and (2) a model adaptation mechanism to learn discriminative features for novel classes while preserving base knowledge using few labeled and all the unlabeled data.
arXiv Detail & Related papers (2021-10-21T13:25:52Z) - HSVA: Hierarchical Semantic-Visual Adaptation for Zero-Shot Learning [74.76431541169342]
Zero-shot learning (ZSL) tackles the unseen class recognition problem, transferring semantic knowledge from seen classes to unseen ones.
We propose a novel hierarchical semantic-visual adaptation (HSVA) framework to align semantic and visual domains.
Experiments on four benchmark datasets demonstrate HSVA achieves superior performance on both conventional and generalized ZSL.
arXiv Detail & Related papers (2021-09-30T14:27:50Z) - A Simple Approach for Zero-Shot Learning based on Triplet Distribution
Embeddings [6.193231258199234]
ZSL aims to recognize unseen classes without labeled training data by exploiting semantic information.
Existing ZSL methods mainly use vectors to represent the embeddings to the semantic space.
We address this issue by leveraging the use of distribution embeddings.
arXiv Detail & Related papers (2021-03-29T20:26:20Z) - Task Aligned Generative Meta-learning for Zero-shot Learning [64.16125851588437]
We propose a Task-aligned Generative Meta-learning model for Zero-shot learning (TGMZ)
TGMZ mitigates the potentially biased training and enables meta-ZSL to accommodate real-world datasets containing diverse distributions.
Our comparisons with state-of-the-art algorithms show the improvements of 2.1%, 3.0%, 2.5%, and 7.6% achieved by TGMZ on AWA1, AWA2, CUB, and aPY datasets.
arXiv Detail & Related papers (2021-03-03T05:18:36Z) - Information Bottleneck Constrained Latent Bidirectional Embedding for
Zero-Shot Learning [59.58381904522967]
We propose a novel embedding based generative model with a tight visual-semantic coupling constraint.
We learn a unified latent space that calibrates the embedded parametric distributions of both visual and semantic spaces.
Our method can be easily extended to transductive ZSL setting by generating labels for unseen images.
arXiv Detail & Related papers (2020-09-16T03:54:12Z) - Generalized Zero-Shot Learning via VAE-Conditioned Generative Flow [83.27681781274406]
Generalized zero-shot learning aims to recognize both seen and unseen classes by transferring knowledge from semantic descriptions to visual representations.
Recent generative methods formulate GZSL as a missing data problem, which mainly adopts GANs or VAEs to generate visual features for unseen classes.
We propose a conditional version of generative flows for GZSL, i.e., VAE-Conditioned Generative Flow (VAE-cFlow)
arXiv Detail & Related papers (2020-09-01T09:12:31Z) - From Anchor Generation to Distribution Alignment: Learning a
Discriminative Embedding Space for Zero-Shot Recognition [46.47620562161315]
In zero-shot learning (ZSL), the samples to be classified are usually projected into side information templates such as attributes.
We propose a novel framework called Discriminative Anchor Generation and Distribution Alignment Model (DAGDA)
Firstly, in order to rectify the distribution of original templates, a diffusion based graph convolutional network, which can explicitly model the interaction between class and side information, is proposed to produce discriminative anchors.
Secondly, to further align the samples with the corresponding anchors in anchor space, which aims to refine the distribution in a fine-grained manner, we introduce a semantic relation regularization
arXiv Detail & Related papers (2020-02-10T05:25:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.