Meta-Learned Attribute Self-Interaction Network for Continual and
Generalized Zero-Shot Learning
- URL: http://arxiv.org/abs/2312.01167v1
- Date: Sat, 2 Dec 2023 16:23:01 GMT
- Title: Meta-Learned Attribute Self-Interaction Network for Continual and
Generalized Zero-Shot Learning
- Authors: Vinay K Verma, Nikhil Mehta, Kevin J Liang, Aakansha Mishra and
Lawrence Carin
- Abstract summary: Zero-shot learning (ZSL) is a promising approach to generalizing a model to unseen categories during training.
We propose a Meta-learned Attribute self-Interaction Network (MAIN) for continual ZSL.
By pairing attribute self-interaction trained using meta-learning with inverse regularization of the attribute encoder, we are able to outperform state-of-the-art results without leveraging the unseen class attributes.
- Score: 46.6282595346048
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Zero-shot learning (ZSL) is a promising approach to generalizing a model to
categories unseen during training by leveraging class attributes, but
challenges remain. Recently, methods using generative models to combat bias
towards classes seen during training have pushed state of the art, but these
generative models can be slow or computationally expensive to train. Also,
these generative models assume that the attribute vector of each unseen class
is available a priori at training, which is not always practical. Additionally,
while many previous ZSL methods assume a one-time adaptation to unseen classes,
in reality, the world is always changing, necessitating a constant adjustment
of deployed models. Models unprepared to handle a sequential stream of data are
likely to experience catastrophic forgetting. We propose a Meta-learned
Attribute self-Interaction Network (MAIN) for continual ZSL. By pairing
attribute self-interaction trained using meta-learning with inverse
regularization of the attribute encoder, we are able to outperform
state-of-the-art results without leveraging the unseen class attributes while
also being able to train our models substantially faster (>100x) than expensive
generative-based approaches. We demonstrate this with experiments on five
standard ZSL datasets (CUB, aPY, AWA1, AWA2, and SUN) in the generalized
zero-shot learning and continual (fixed/dynamic) zero-shot learning settings.
Extensive ablations and analyses demonstrate the efficacy of various components
proposed.
Related papers
- CLIPood: Generalizing CLIP to Out-of-Distributions [73.86353105017076]
Contrastive language-image pre-training (CLIP) models have shown impressive zero-shot ability, but the further adaptation of CLIP on downstream tasks undesirably degrades OOD performances.
We propose CLIPood, a fine-tuning method that can adapt CLIP models to OOD situations where both domain shifts and open classes may occur on unseen test data.
Experiments on diverse datasets with different OOD scenarios show that CLIPood consistently outperforms existing generalization techniques.
arXiv Detail & Related papers (2023-02-02T04:27:54Z) - Efficient Gaussian Process Model on Class-Imbalanced Datasets for
Generalized Zero-Shot Learning [37.00463358780726]
We propose a Neural Network model that learns a latent feature embedding and a Gaussian Process (GP) regression model that predicts latent feature prototypes of unseen classes.
Our model is trained efficiently with a simple training strategy that mitigates the impact of class-imbalanced training data.
arXiv Detail & Related papers (2022-10-11T04:57:20Z) - Adaptive Memory Networks with Self-supervised Learning for Unsupervised
Anomaly Detection [54.76993389109327]
Unsupervised anomaly detection aims to build models to detect unseen anomalies by only training on the normal data.
We propose a novel approach called Adaptive Memory Network with Self-supervised Learning (AMSL) to address these challenges.
AMSL incorporates a self-supervised learning module to learn general normal patterns and an adaptive memory fusion module to learn rich feature representations.
arXiv Detail & Related papers (2022-01-03T03:40:21Z) - Dynamic VAEs with Generative Replay for Continual Zero-shot Learning [1.90365714903665]
This paper proposes a novel continual zero-shot learning (DVGR-CZSL) model that grows in size with each task and uses generative replay to update itself with previously learned classes to avoid forgetting.
We show our method is superior in task sequentially learning with ZSL(Zero-Shot Learning)
arXiv Detail & Related papers (2021-04-26T10:56:43Z) - Attribute-Modulated Generative Meta Learning for Zero-Shot
Classification [52.64680991682722]
We present the Attribute-Modulated generAtive meta-model for Zero-shot learning (AMAZ)
Our model consists of an attribute-aware modulation network and an attribute-augmented generative network.
Our empirical evaluations show that AMAZ improves state-of-the-art methods by 3.8% and 5.1% in ZSL and generalized ZSL settings, respectively.
arXiv Detail & Related papers (2021-04-22T04:16:43Z) - Meta-Learned Attribute Self-Gating for Continual Generalized Zero-Shot
Learning [82.07273754143547]
We propose a meta-continual zero-shot learning (MCZSL) approach to generalizing a model to categories unseen during training.
By pairing self-gating of attributes and scaled class normalization with meta-learning based training, we are able to outperform state-of-the-art results.
arXiv Detail & Related papers (2021-02-23T18:36:14Z) - CLASTER: Clustering with Reinforcement Learning for Zero-Shot Action
Recognition [52.66360172784038]
We propose a clustering-based model, which considers all training samples at once, instead of optimizing for each instance individually.
We call the proposed method CLASTER and observe that it consistently improves over the state-of-the-art in all standard datasets.
arXiv Detail & Related papers (2021-01-18T12:46:24Z) - Towards Zero-Shot Learning with Fewer Seen Class Examples [41.751885300474925]
We present a meta-learning based generative model for zero-shot learning (ZSL)
This setup contrasts with the conventional ZSL approaches, where training typically assumes the availability of a sufficiently large number of training examples from each of the seen classes.
We conduct extensive experiments and ablation studies on four benchmark datasets of ZSL and observe that the proposed model outperforms state-of-the-art approaches by a significant margin when the number of examples per seen class is very small.
arXiv Detail & Related papers (2020-11-14T11:58:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.