Targeted Attention for Generalized- and Zero-Shot Learning
- URL: http://arxiv.org/abs/2211.09322v1
- Date: Thu, 17 Nov 2022 03:55:18 GMT
- Title: Targeted Attention for Generalized- and Zero-Shot Learning
- Authors: Abhijit Suprem
- Abstract summary: The Zero-Shot Learning (ZSL) task attempts to learn concepts without any labeled data.
We show state-of-the-art results in the Generalized Zero-Shot Learning (GZSL) setting, with Harmonic Mean R-1 of 66.14% on the CUB200 dataset.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Zero-Shot Learning (ZSL) task attempts to learn concepts without any
labeled data. Unlike traditional classification/detection tasks, the evaluation
environment is provided unseen classes never encountered during training. As
such, it remains both challenging, and promising on a variety of fronts,
including unsupervised concept learning, domain adaptation, and dataset drift
detection. Recently, there have been a variety of approaches towards solving
ZSL, including improved metric learning methods, transfer learning,
combinations of semantic and image domains using, e.g. word vectors, and
generative models to model the latent space of known classes to classify unseen
classes. We find many approaches require intensive training augmentation with
attributes or features that may be commonly unavailable (attribute-based
learning) or susceptible to adversarial attacks (generative learning). We
propose combining approaches from the related person re-identification task for
ZSL, with key modifications to ensure sufficiently improved performance in the
ZSL setting without the need for feature or training dataset augmentation. We
are able to achieve state-of-the-art performance on the CUB200 and Cars196
datasets in the ZSL setting compared to recent works, with NMI (normalized
mutual inference) of 63.27 and top-1 of 61.04 for CUB200, and NMI 66.03 with
top-1 82.75% in Cars196. We also show state-of-the-art results in the
Generalized Zero-Shot Learning (GZSL) setting, with Harmonic Mean R-1 of 66.14%
on the CUB200 dataset.
Related papers
- Meta-Learned Attribute Self-Interaction Network for Continual and
Generalized Zero-Shot Learning [46.6282595346048]
Zero-shot learning (ZSL) is a promising approach to generalizing a model to unseen categories during training.
We propose a Meta-learned Attribute self-Interaction Network (MAIN) for continual ZSL.
By pairing attribute self-interaction trained using meta-learning with inverse regularization of the attribute encoder, we are able to outperform state-of-the-art results without leveraging the unseen class attributes.
arXiv Detail & Related papers (2023-12-02T16:23:01Z) - SLCA: Slow Learner with Classifier Alignment for Continual Learning on a
Pre-trained Model [73.80068155830708]
We present an extensive analysis for continual learning on a pre-trained model (CLPM)
We propose a simple but extremely effective approach named Slow Learner with Alignment (SLCA)
Across a variety of scenarios, our proposal provides substantial improvements for CLPM.
arXiv Detail & Related papers (2023-03-09T08:57:01Z) - DiRaC-I: Identifying Diverse and Rare Training Classes for Zero-Shot
Learning [6.75714270653184]
It is intuitive that intelligently selecting the training classes from a dataset for Zero-Shot Learning (ZSL) can improve the performance of existing ZSL methods.
We propose a framework called Diverse and Rare Class Identifier (DiRaC-I) which, given an attribute-based dataset, can intelligently yield the most suitable "seen classes" for training ZSL models.
Our results demonstrate DiRaC-I helps ZSL models to achieve significant classification accuracy improvements.
arXiv Detail & Related papers (2022-12-31T16:05:09Z) - Federated Zero-Shot Learning for Visual Recognition [55.65879596326147]
We propose a novel Federated Zero-Shot Learning FedZSL framework.
FedZSL learns a central model from the decentralized data residing on edge devices.
The effectiveness and robustness of FedZSL are demonstrated by extensive experiments conducted on three zero-shot benchmark datasets.
arXiv Detail & Related papers (2022-09-05T14:49:34Z) - Attribute-Modulated Generative Meta Learning for Zero-Shot
Classification [52.64680991682722]
We present the Attribute-Modulated generAtive meta-model for Zero-shot learning (AMAZ)
Our model consists of an attribute-aware modulation network and an attribute-augmented generative network.
Our empirical evaluations show that AMAZ improves state-of-the-art methods by 3.8% and 5.1% in ZSL and generalized ZSL settings, respectively.
arXiv Detail & Related papers (2021-04-22T04:16:43Z) - Task Aligned Generative Meta-learning for Zero-shot Learning [64.16125851588437]
We propose a Task-aligned Generative Meta-learning model for Zero-shot learning (TGMZ)
TGMZ mitigates the potentially biased training and enables meta-ZSL to accommodate real-world datasets containing diverse distributions.
Our comparisons with state-of-the-art algorithms show the improvements of 2.1%, 3.0%, 2.5%, and 7.6% achieved by TGMZ on AWA1, AWA2, CUB, and aPY datasets.
arXiv Detail & Related papers (2021-03-03T05:18:36Z) - Meta-Learned Attribute Self-Gating for Continual Generalized Zero-Shot
Learning [82.07273754143547]
We propose a meta-continual zero-shot learning (MCZSL) approach to generalizing a model to categories unseen during training.
By pairing self-gating of attributes and scaled class normalization with meta-learning based training, we are able to outperform state-of-the-art results.
arXiv Detail & Related papers (2021-02-23T18:36:14Z) - OntoZSL: Ontology-enhanced Zero-shot Learning [19.87808305218359]
Key to implementing Zero-shot Learning (ZSL) is to leverage the prior knowledge of classes which builds the semantic relationship between classes.
In this paper, we explore richer and more competitive prior knowledge to model the inter-class relationship for ZSL.
To address the data imbalance between seen classes and unseen classes, we developed a generative ZSL framework with Generative Adversarial Networks (GANs)
arXiv Detail & Related papers (2021-02-15T04:39:58Z) - CLASTER: Clustering with Reinforcement Learning for Zero-Shot Action
Recognition [52.66360172784038]
We propose a clustering-based model, which considers all training samples at once, instead of optimizing for each instance individually.
We call the proposed method CLASTER and observe that it consistently improves over the state-of-the-art in all standard datasets.
arXiv Detail & Related papers (2021-01-18T12:46:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.