Simple and effective localized attribute representations for zero-shot
learning
- URL: http://arxiv.org/abs/2006.05938v3
- Date: Tue, 9 Mar 2021 09:44:15 GMT
- Title: Simple and effective localized attribute representations for zero-shot
learning
- Authors: Shiqi Yang, Kai Wang, Luis Herranz, Joost van de Weijer
- Abstract summary: Zero-shot learning (ZSL) aims to discriminate images from unseen classes by exploiting relations to seen classes via their semantic descriptions.
We propose localizing representations in the semantic/attribute space, with a simple but effective pipeline where localization is implicit.
Our method can be implemented easily, which can be used as a new baseline for zero shot-learning.
- Score: 48.053204004771665
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Zero-shot learning (ZSL) aims to discriminate images from unseen classes by
exploiting relations to seen classes via their semantic descriptions. Some
recent papers have shown the importance of localized features together with
fine-tuning the feature extractor to obtain discriminative and transferable
features. However, these methods require complex attention or part detection
modules to perform explicit localization in the visual space. In contrast, in
this paper we propose localizing representations in the semantic/attribute
space, with a simple but effective pipeline where localization is implicit.
Focusing on attribute representations, we show that our method obtains
state-of-the-art performance on CUB and SUN datasets, and also achieves
competitive results on AWA2 dataset, outperforming generally more complex
methods with explicit localization in the visual space. Our method can be
implemented easily, which can be used as a new baseline for zero shot-learning.
In addition, our localized representations are highly interpretable as
attribute-specific heatmaps.
Related papers
- Attribute-Aware Representation Rectification for Generalized Zero-Shot
Learning [19.65026043141699]
Generalized Zero-shot Learning (GZSL) has yielded remarkable performance by designing a series of unbiased visual-semantics mappings.
We propose a simple yet effective Attribute-Aware Representation Rectification framework for GZSL, dubbed $mathbf(AR)2$.
arXiv Detail & Related papers (2023-11-23T11:30:32Z) - Hierarchical Visual Primitive Experts for Compositional Zero-Shot
Learning [52.506434446439776]
Compositional zero-shot learning (CZSL) aims to recognize compositions with prior knowledge of known primitives (attribute and object)
We propose a simple and scalable framework called Composition Transformer (CoT) to address these issues.
Our method achieves SoTA performance on several benchmarks, including MIT-States, C-GQA, and VAW-CZSL.
arXiv Detail & Related papers (2023-08-08T03:24:21Z) - Towards Effective Image Manipulation Detection with Proposal Contrastive
Learning [61.5469708038966]
We propose Proposal Contrastive Learning (PCL) for effective image manipulation detection.
Our PCL consists of a two-stream architecture by extracting two types of global features from RGB and noise views respectively.
Our PCL can be easily adapted to unlabeled data in practice, which can reduce manual labeling costs and promote more generalizable features.
arXiv Detail & Related papers (2022-10-16T13:30:13Z) - Attribute Prototype Network for Any-Shot Learning [113.50220968583353]
We argue that an image representation with integrated attribute localization ability would be beneficial for any-shot, i.e. zero-shot and few-shot, image classification tasks.
We propose a novel representation learning framework that jointly learns global and local features using only class-level attributes.
arXiv Detail & Related papers (2022-04-04T02:25:40Z) - Region Semantically Aligned Network for Zero-Shot Learning [18.18665627472823]
We propose a Region Semantically Aligned Network (RSAN) which maps local features of unseen classes to their semantic attributes.
We obtain each attribute from a specific region of the output and exploit these attributes for recognition.
Experiments on several standard ZSL datasets reveal the benefit of the proposed RSAN method, outperforming state-of-the-art methods.
arXiv Detail & Related papers (2021-10-14T03:23:40Z) - On Implicit Attribute Localization for Generalized Zero-Shot Learning [43.61533666141709]
We show that common ZSL backbones can implicitly localize attributes, yet this property is not exploited.
We then propose SELAR, a simple method that further encourages attribute localization, surprisingly achieving very competitive generalized ZSL (GZSL) performance.
arXiv Detail & Related papers (2021-03-08T12:31:37Z) - Goal-Oriented Gaze Estimation for Zero-Shot Learning [62.52340838817908]
We introduce a novel goal-oriented gaze estimation module (GEM) to improve the discriminative attribute localization.
We aim to predict the actual human gaze location to get the visual attention regions for recognizing a novel object guided by attribute description.
This work implies the promising benefits of collecting human gaze dataset and automatic gaze estimation algorithms on high-level computer vision tasks.
arXiv Detail & Related papers (2021-03-05T02:14:57Z) - Attribute Prototype Network for Zero-Shot Learning [113.50220968583353]
We propose a novel zero-shot representation learning framework that jointly learns discriminative global and local features.
Our model points to the visual evidence of the attributes in an image, confirming the improved attribute localization ability of our image representation.
arXiv Detail & Related papers (2020-08-19T06:46:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.