Vocabulary-informed Zero-shot and Open-set Learning
- URL: http://arxiv.org/abs/2301.00998v2
- Date: Wed, 4 Jan 2023 02:09:33 GMT
- Title: Vocabulary-informed Zero-shot and Open-set Learning
- Authors: Yanwei Fu, Xiaomei Wang, Hanze Dong, Yu-Gang Jiang, Meng Wang,
Xiangyang Xue, Leonid Sigal
- Abstract summary: We propose vocabulary-informed learning to address problems of supervised, zero-shot, generalized zero-shot and open set recognition.
Specifically, we propose a weighted maximum margin framework for semantic manifold-based recognition that incorporates distance constraints from (both supervised and unsupervised) vocabulary atoms.
We illustrate that resulting model shows improvements in supervised, zero-shot, generalized zero-shot, and large open set recognition, with up to 310K class vocabulary on Animal with Attributes and ImageNet datasets.
- Score: 128.83517181045815
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite significant progress in object categorization, in recent years, a
number of important challenges remain; mainly, the ability to learn from
limited labeled data and to recognize object classes within large, potentially
open, set of labels. Zero-shot learning is one way of addressing these
challenges, but it has only been shown to work with limited sized class
vocabularies and typically requires separation between supervised and
unsupervised classes, allowing former to inform the latter but not vice versa.
We propose the notion of vocabulary-informed learning to alleviate the above
mentioned challenges and address problems of supervised, zero-shot, generalized
zero-shot and open set recognition using a unified framework. Specifically, we
propose a weighted maximum margin framework for semantic manifold-based
recognition that incorporates distance constraints from (both supervised and
unsupervised) vocabulary atoms. Distance constraints ensure that labeled
samples are projected closer to their correct prototypes, in the embedding
space, than to others. We illustrate that resulting model shows improvements in
supervised, zero-shot, generalized zero-shot, and large open set recognition,
with up to 310K class vocabulary on Animal with Attributes and ImageNet
datasets.
Related papers
- Towards Realistic Zero-Shot Classification via Self Structural Semantic
Alignment [53.2701026843921]
Large-scale pre-trained Vision Language Models (VLMs) have proven effective for zero-shot classification.
In this paper, we aim at a more challenging setting, Realistic Zero-Shot Classification, which assumes no annotation but instead a broad vocabulary.
We propose the Self Structural Semantic Alignment (S3A) framework, which extracts structural semantic information from unlabeled data while simultaneously self-learning.
arXiv Detail & Related papers (2023-08-24T17:56:46Z) - Towards Open Vocabulary Learning: A Survey [146.90188069113213]
Deep neural networks have made impressive advancements in various core tasks like segmentation, tracking, and detection.
Recently, open vocabulary settings were proposed due to the rapid progress of vision language pre-training.
This paper provides a thorough review of open vocabulary learning, summarizing and analyzing recent developments in the field.
arXiv Detail & Related papers (2023-06-28T02:33:06Z) - Simple Linguistic Inferences of Large Language Models (LLMs): Blind Spots and Blinds [59.71218039095155]
We evaluate language understanding capacities on simple inference tasks that most humans find trivial.
We target (i) grammatically-specified entailments, (ii) premises with evidential adverbs of uncertainty, and (iii) monotonicity entailments.
The models exhibit moderate to low performance on these evaluation sets.
arXiv Detail & Related papers (2023-05-24T06:41:09Z) - Zero-Shot Learning for Joint Intent and Slot Labeling [11.82805641934772]
We show that one can profitably perform joint zero-shot intent classification and slot labeling.
We describe NN architectures that translate between word and sentence embedding spaces.
arXiv Detail & Related papers (2022-11-29T01:58:25Z) - Zero-Shot Text Classification with Self-Training [8.68603153534916]
We show that fine-tuning the zero-shot classifier on its most confident predictions leads to significant performance gains across a wide range of text classification tasks.
Self-training adapts the zero-shot model to the task at hand.
arXiv Detail & Related papers (2022-10-31T17:55:00Z) - Analyzing the Limits of Self-Supervision in Handling Bias in Language [52.26068057260399]
We evaluate how well language models capture the semantics of four tasks for bias: diagnosis, identification, extraction and rephrasing.
Our analyses indicate that language models are capable of performing these tasks to widely varying degrees across different bias dimensions, such as gender and political affiliation.
arXiv Detail & Related papers (2021-12-16T05:36:08Z) - CLASTER: Clustering with Reinforcement Learning for Zero-Shot Action
Recognition [52.66360172784038]
We propose a clustering-based model, which considers all training samples at once, instead of optimizing for each instance individually.
We call the proposed method CLASTER and observe that it consistently improves over the state-of-the-art in all standard datasets.
arXiv Detail & Related papers (2021-01-18T12:46:24Z) - Linguistically-Enriched and Context-Aware Zero-shot Slot Filling [6.06746295810681]
Slot filling is one of the most important challenges in modern task-oriented dialog systems.
New domains (i.e., unseen in training) may emerge after deployment.
It is imperative that models seamlessly adapt and fill slots from both seen and unseen domains.
arXiv Detail & Related papers (2021-01-16T20:18:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.