Attribute Propagation Network for Graph Zero-shot Learning
- URL: http://arxiv.org/abs/2009.11816v1
- Date: Thu, 24 Sep 2020 16:53:40 GMT
- Title: Attribute Propagation Network for Graph Zero-shot Learning
- Authors: Lu Liu, Tianyi Zhou, Guodong Long, Jing Jiang, Chengqi Zhang
- Abstract summary: We introduce the attribute propagation network (APNet), which is composed of 1) a graph propagation model generating attribute vector for each class and 2) a parameterized nearest neighbor (NN) classifier.
APNet achieves either compelling performance or new state-of-the-art results in experiments with two zero-shot learning settings and five benchmark datasets.
- Score: 57.68486382473194
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The goal of zero-shot learning (ZSL) is to train a model to classify samples
of classes that were not seen during training. To address this challenging
task, most ZSL methods relate unseen test classes to seen(training) classes via
a pre-defined set of attributes that can describe all classes in the same
semantic space, so the knowledge learned on the training classes can be adapted
to unseen classes. In this paper, we aim to optimize the attribute space for
ZSL by training a propagation mechanism to refine the semantic attributes of
each class based on its neighbors and related classes on a graph of classes. We
show that the propagated attributes can produce classifiers for zero-shot
classes with significantly improved performance in different ZSL settings. The
graph of classes is usually free or very cheap to acquire such as WordNet or
ImageNet classes. When the graph is not provided, given pre-defined semantic
embeddings of the classes, we can learn a mechanism to generate the graph in an
end-to-end manner along with the propagation mechanism. However, this
graph-aided technique has not been well-explored in the literature. In this
paper, we introduce the attribute propagation network (APNet), which is
composed of 1) a graph propagation model generating attribute vector for each
class and 2) a parameterized nearest neighbor (NN) classifier categorizing an
image to the class with the nearest attribute vector to the image's embedding.
For better generalization over unseen classes, different from previous methods,
we adopt a meta-learning strategy to train the propagation mechanism and the
similarity metric for the NN classifier on multiple sub-graphs, each associated
with a classification task over a subset of training classes. In experiments
with two zero-shot learning settings and five benchmark datasets, APNet
achieves either compelling performance or new state-of-the-art results.
Related papers
- Text Descriptions are Compressive and Invariant Representations for
Visual Learning [63.3464863723631]
We show that an alternative approach, in line with humans' understanding of multiple visual features per class, can provide compelling performance in the robust few-shot learning setting.
In particular, we introduce a novel method, textit SLR-AVD (Sparse Logistic Regression using Augmented Visual Descriptors).
This method first automatically generates multiple visual descriptions of each class via a large language model (LLM), then uses a VLM to translate these descriptions to a set of visual feature embeddings of each image, and finally uses sparse logistic regression to select a relevant subset of these features to classify
arXiv Detail & Related papers (2023-07-10T03:06:45Z) - APANet: Adaptive Prototypes Alignment Network for Few-Shot Semantic
Segmentation [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a given query image with only a few labeled support images.
Most advanced solutions exploit a metric learning framework that performs segmentation through matching each query feature to a learned class-specific prototype.
We present an adaptive prototype representation by introducing class-specific and class-agnostic prototypes.
arXiv Detail & Related papers (2021-11-24T04:38:37Z) - GAN for Vision, KG for Relation: a Two-stage Deep Network for Zero-shot
Action Recognition [33.23662792742078]
We propose a two-stage deep neural network for zero-shot action recognition.
In the sampling stage, we utilize a generative adversarial networks (GAN) trained by action features and word vectors of seen classes.
In the classification stage, we construct a knowledge graph based on the relationship between word vectors of action classes and related objects.
arXiv Detail & Related papers (2021-05-25T09:34:42Z) - OntoZSL: Ontology-enhanced Zero-shot Learning [19.87808305218359]
Key to implementing Zero-shot Learning (ZSL) is to leverage the prior knowledge of classes which builds the semantic relationship between classes.
In this paper, we explore richer and more competitive prior knowledge to model the inter-class relationship for ZSL.
To address the data imbalance between seen classes and unseen classes, we developed a generative ZSL framework with Generative Adversarial Networks (GANs)
arXiv Detail & Related papers (2021-02-15T04:39:58Z) - Isometric Propagation Network for Generalized Zero-shot Learning [72.02404519815663]
A popular strategy is to learn a mapping between the semantic space of class attributes and the visual space of images based on the seen classes and their data.
We propose Isometric propagation Network (IPN), which learns to strengthen the relation between classes within each space and align the class dependency in the two spaces.
IPN achieves state-of-the-art performance on three popular Zero-shot learning benchmarks.
arXiv Detail & Related papers (2021-02-03T12:45:38Z) - Generative Multi-Label Zero-Shot Learning [136.17594611722285]
Multi-label zero-shot learning strives to classify images into multiple unseen categories for which no data is available during training.
Our work is the first to tackle the problem of multi-label feature in the (generalized) zero-shot setting.
Our cross-level fusion-based generative approach outperforms the state-of-the-art on all three datasets.
arXiv Detail & Related papers (2021-01-27T18:56:46Z) - CLASTER: Clustering with Reinforcement Learning for Zero-Shot Action
Recognition [52.66360172784038]
We propose a clustering-based model, which considers all training samples at once, instead of optimizing for each instance individually.
We call the proposed method CLASTER and observe that it consistently improves over the state-of-the-art in all standard datasets.
arXiv Detail & Related papers (2021-01-18T12:46:24Z) - A Multi-class Approach -- Building a Visual Classifier based on Textual
Descriptions using Zero-Shot Learning [0.34265828682659694]
We overcome the two main hurdles of Machine Learning, i.e. scarcity of data and constrained prediction of the classification model.
We train a classifier by mapping labelled images to their textual description instead of training it for specific classes.
arXiv Detail & Related papers (2020-11-18T12:06:55Z) - Generative Adversarial Zero-shot Learning via Knowledge Graphs [32.42721467499858]
We introduce a new generative ZSL method named KG-GAN by incorporating rich semantics in a knowledge graph (KG) into GANs.
Specifically, we build upon Graph Neural Networks and encode KG from two views: class view and attribute view.
With well-learned semantic embeddings for each node (representing a visual category), we leverage GANs to synthesize compelling visual features for unseen classes.
arXiv Detail & Related papers (2020-04-07T03:55:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.