Towards Unbiased Multi-label Zero-Shot Learning with Pyramid and
Semantic Attention
- URL: http://arxiv.org/abs/2203.03483v1
- Date: Mon, 7 Mar 2022 15:52:46 GMT
- Title: Towards Unbiased Multi-label Zero-Shot Learning with Pyramid and
Semantic Attention
- Authors: Ziming Liu, Song Guo, Jingcai Guo, Yuanyuan Xu, Fushuo Huo
- Abstract summary: Multi-label zero-shot learning aims at recognizing multiple unseen labels of classes for each input sample.
We propose a novel framework of unbiased multi-label zero-shot learning, by considering various class-specific regions.
- Score: 14.855116554722489
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-label zero-shot learning extends conventional single-label zero-shot
learning to a more realistic scenario that aims at recognizing multiple unseen
labels of classes for each input sample. Existing works usually exploit
attention mechanism to generate the correlation among different labels.
However, most of them are usually biased on several major classes while neglect
most of the minor classes with the same importance in input samples, and may
thus result in overly diffused attention maps that cannot sufficiently cover
minor classes. We argue that disregarding the connection between major and
minor classes, i.e., correspond to the global and local information,
respectively, is the cause of the problem. In this paper, we propose a novel
framework of unbiased multi-label zero-shot learning, by considering various
class-specific regions to calibrate the training process of the classifier.
Specifically, Pyramid Feature Attention (PFA) is proposed to build the
correlation between global and local information of samples to balance the
presence of each class. Meanwhile, for the generated semantic representations
of input samples, we propose Semantic Attention (SA) to strengthen the
element-wise correlation among these vectors, which can encourage the
coordinated representation of them. Extensive experiments on the large-scale
multi-label zero-shot benchmarks NUS-WIDE and Open-Image demonstrate that the
proposed method surpasses other representative methods by significant margins.
Related papers
- Beyond the Known: Novel Class Discovery for Open-world Graph Learning [16.30962452905747]
We propose an Open-world gRAph neuraL network (ORAL) to tackle novel class discovery on graphs.
ORAL first detects correlations between classes through semi-supervised prototypical learning.
To fully explore multi-scale graph features for alleviating label deficiencies, ORAL generates pseudo-labels.
arXiv Detail & Related papers (2024-03-29T01:25:05Z) - Exploring Vacant Classes in Label-Skewed Federated Learning [113.65301899666645]
Label skews, characterized by disparities in local label distribution across clients, pose a significant challenge in federated learning.
This paper introduces FedVLS, a novel approach to label-skewed federated learning that integrates vacant-class distillation and logit suppression simultaneously.
arXiv Detail & Related papers (2024-01-04T16:06:31Z) - Virtual Category Learning: A Semi-Supervised Learning Method for Dense
Prediction with Extremely Limited Labels [63.16824565919966]
This paper proposes to use confusing samples proactively without label correction.
A Virtual Category (VC) is assigned to each confusing sample in such a way that it can safely contribute to the model optimisation.
Our intriguing findings highlight the usage of VC learning in dense vision tasks.
arXiv Detail & Related papers (2023-12-02T16:23:52Z) - Evolving Multi-Label Fuzzy Classifier [5.53329677986653]
Multi-label classification has attracted much attention in the machine learning community to address the problem of assigning single samples to more than one class at the same time.
We propose an evolving multi-label fuzzy classifier (EFC-ML) which is able to self-adapt and self-evolve its structure with new incoming multi-label samples in an incremental, single-pass manner.
arXiv Detail & Related papers (2022-03-29T08:01:03Z) - Self-Training: A Survey [5.772546394254112]
Semi-supervised algorithms aim to learn prediction functions from a small set of labeled observations and a large set of unlabeled observations.
Among the existing techniques, self-training methods have undoubtedly attracted greater attention in recent years.
We present self-training methods for binary and multi-class classification; as well as their variants and two related approaches.
arXiv Detail & Related papers (2022-02-24T11:40:44Z) - Discriminative Region-based Multi-Label Zero-Shot Learning [145.0952336375342]
Multi-label zero-shot learning (ZSL) is a more realistic counter-part of standard single-label ZSL.
We propose an alternate approach towards region-based discriminability-preserving ZSL.
arXiv Detail & Related papers (2021-08-20T17:56:47Z) - Generative Multi-Label Zero-Shot Learning [136.17594611722285]
Multi-label zero-shot learning strives to classify images into multiple unseen categories for which no data is available during training.
Our work is the first to tackle the problem of multi-label feature in the (generalized) zero-shot setting.
Our cross-level fusion-based generative approach outperforms the state-of-the-art on all three datasets.
arXiv Detail & Related papers (2021-01-27T18:56:46Z) - CLASTER: Clustering with Reinforcement Learning for Zero-Shot Action
Recognition [52.66360172784038]
We propose a clustering-based model, which considers all training samples at once, instead of optimizing for each instance individually.
We call the proposed method CLASTER and observe that it consistently improves over the state-of-the-art in all standard datasets.
arXiv Detail & Related papers (2021-01-18T12:46:24Z) - Towards Cross-Granularity Few-Shot Learning: Coarse-to-Fine
Pseudo-Labeling with Visual-Semantic Meta-Embedding [13.063136901934865]
Few-shot learning aims at rapidly adapting to novel categories with only a handful of samples at test time.
In this paper, we advance the few-shot classification paradigm towards a more challenging scenario, i.e., cross-granularity few-shot classification.
We approximate the fine-grained data distribution by greedy clustering of each coarse-class into pseudo-fine-classes according to the similarity of image embeddings.
arXiv Detail & Related papers (2020-07-11T03:44:21Z) - Federated Learning with Only Positive Labels [71.63836379169315]
We propose a generic framework for training with only positive labels, namely Federated Averaging with Spreadout (FedAwS)
We show, both theoretically and empirically, that FedAwS can almost match the performance of conventional learning where users have access to negative labels.
arXiv Detail & Related papers (2020-04-21T23:35:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.