Boosting Generative Zero-Shot Learning by Synthesizing Diverse Features
with Attribute Augmentation
- URL: http://arxiv.org/abs/2112.12573v1
- Date: Thu, 23 Dec 2021 14:32:51 GMT
- Title: Boosting Generative Zero-Shot Learning by Synthesizing Diverse Features
with Attribute Augmentation
- Authors: Xiaojie Zhao, Yuming Shen, Shidong Wang, Haofeng Zhang
- Abstract summary: We propose a novel framework to boost Zero-Shot Learning (ZSL) by synthesizing diverse features.
This method uses augmented semantic attributes to train the generative model, so as to simulate the real distribution of visual features.
We evaluate the proposed model on four benchmark datasets, observing significant performance improvement against the state-of-the-art.
- Score: 21.72622601533585
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recent advance in deep generative models outlines a promising perspective
in the realm of Zero-Shot Learning (ZSL). Most generative ZSL methods use
category semantic attributes plus a Gaussian noise to generate visual features.
After generating unseen samples, this family of approaches effectively
transforms the ZSL problem into a supervised classification scheme. However,
the existing models use a single semantic attribute, which contains the
complete attribute information of the category. The generated data also carry
the complete attribute information, but in reality, visual samples usually have
limited attributes. Therefore, the generated data from attribute could have
incomplete semantics. Based on this fact, we propose a novel framework to boost
ZSL by synthesizing diverse features. This method uses augmented semantic
attributes to train the generative model, so as to simulate the real
distribution of visual features. We evaluate the proposed model on four
benchmark datasets, observing significant performance improvement against the
state-of-the-art.
Related papers
- Visual-Augmented Dynamic Semantic Prototype for Generative Zero-Shot Learning [56.16593809016167]
We propose a novel Visual-Augmented Dynamic Semantic prototype method (termed VADS) to boost the generator to learn accurate semantic-visual mapping.
VADS consists of two modules: (1) Visual-aware Domain Knowledge Learning module (VDKL) learns the local bias and global prior of the visual features, which replace pure Gaussian noise to provide richer prior noise information; (2) Vision-Oriented Semantic Updation module (VOSU) updates the semantic prototype according to the visual representations of the samples.
arXiv Detail & Related papers (2024-04-23T07:39:09Z) - Dual Feature Augmentation Network for Generalized Zero-shot Learning [14.410978100610489]
Zero-shot learning (ZSL) aims to infer novel classes without training samples by transferring knowledge from seen classes.
Existing embedding-based approaches for ZSL typically employ attention mechanisms to locate attributes on an image.
We propose a novel Dual Feature Augmentation Network (DFAN), which comprises two feature augmentation modules.
arXiv Detail & Related papers (2023-09-25T02:37:52Z) - Hierarchical Visual Primitive Experts for Compositional Zero-Shot
Learning [52.506434446439776]
Compositional zero-shot learning (CZSL) aims to recognize compositions with prior knowledge of known primitives (attribute and object)
We propose a simple and scalable framework called Composition Transformer (CoT) to address these issues.
Our method achieves SoTA performance on several benchmarks, including MIT-States, C-GQA, and VAW-CZSL.
arXiv Detail & Related papers (2023-08-08T03:24:21Z) - Evolving Semantic Prototype Improves Generative Zero-Shot Learning [73.07035277030573]
In zero-shot learning (ZSL), generative methods synthesize class-related sample features based on predefined semantic prototypes.
We observe that each class's predefined semantic prototype does not accurately match its real semantic prototype.
We propose a dynamic semantic prototype evolving (DSP) method to align the empirically predefined semantic prototypes and the real prototypes for class-related feature synthesis.
arXiv Detail & Related papers (2023-06-12T08:11:06Z) - Exploiting Semantic Attributes for Transductive Zero-Shot Learning [97.61371730534258]
Zero-shot learning aims to recognize unseen classes by generalizing the relation between visual features and semantic attributes learned from the seen classes.
We present a novel transductive ZSL method that produces semantic attributes of the unseen data and imposes them on the generative process.
Experiments on five standard benchmarks show that our method yields state-of-the-art results for zero-shot learning.
arXiv Detail & Related papers (2023-03-17T09:09:48Z) - Boosting Zero-shot Learning via Contrastive Optimization of Attribute
Representations [28.46906100680767]
We propose a new framework to boost Zero-shot learning (ZSL) by explicitly learning attribute prototypes beyond images.
A prototype generation module is designed to generate attribute prototypes from attribute semantics.
A hard example-based contrastive optimization scheme is introduced to reinforce attribute-level features in the embedding space.
arXiv Detail & Related papers (2022-07-08T11:05:35Z) - GSMFlow: Generation Shifts Mitigating Flow for Generalized Zero-Shot
Learning [55.79997930181418]
Generalized Zero-Shot Learning aims to recognize images from both the seen and unseen classes by transferring semantic knowledge from seen to unseen classes.
It is a promising solution to take the advantage of generative models to hallucinate realistic unseen samples based on the knowledge learned from the seen classes.
We propose a novel flow-based generative framework that consists of multiple conditional affine coupling layers for learning unseen data generation.
arXiv Detail & Related papers (2022-07-05T04:04:37Z) - Attribute-Modulated Generative Meta Learning for Zero-Shot
Classification [52.64680991682722]
We present the Attribute-Modulated generAtive meta-model for Zero-shot learning (AMAZ)
Our model consists of an attribute-aware modulation network and an attribute-augmented generative network.
Our empirical evaluations show that AMAZ improves state-of-the-art methods by 3.8% and 5.1% in ZSL and generalized ZSL settings, respectively.
arXiv Detail & Related papers (2021-04-22T04:16:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.