Open World Compositional Zero-Shot Learning
- URL: http://arxiv.org/abs/2101.12609v1
- Date: Fri, 29 Jan 2021 14:45:52 GMT
- Title: Open World Compositional Zero-Shot Learning
- Authors: Massimiliano Mancini, Muhammad Ferjad Naeem, Yongqin Xian, Zeynep
Akata
- Abstract summary: Compositional Zero-Shot learning (CZSL) requires to recognize state-object compositions unseen during training.
We operate on the open world setting, where the search space includes a large number of unseen compositions.
- Score: 47.09665742252187
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Compositional Zero-Shot learning (CZSL) requires to recognize state-object
compositions unseen during training. In this work, instead of assuming the
presence of prior knowledge about the unseen compositions, we operate on the
open world setting, where the search space includes a large number of unseen
compositions some of which might be unfeasible. In this setting, we start from
the cosine similarity between visual features and compositional embeddings.
After estimating the feasibility score of each composition, we use these scores
to either directly mask the output space or as a margin for the cosine
similarity between visual features and compositional embeddings during
training. Our experiments on two standard CZSL benchmarks show that all the
methods suffer severe performance degradation when applied in the open world
setting. While our simple CZSL model achieves state-of-the-art performances in
the closed world scenario, our feasibility scores boost the performance of our
approach in the open world setting, clearly outperforming the previous state of
the art.
Related papers
- Cross-composition Feature Disentanglement for Compositional Zero-shot Learning [49.919635694894204]
Disentanglement of visual features of primitives (i.e., attributes and objects) has shown exceptional results in Compositional Zero-shot Learning (CZSL)
We propose the solution of cross-composition feature disentanglement, which takes multiple primitive-sharing compositions as inputs and constrains the disentangled primitive features to be general across these compositions.
arXiv Detail & Related papers (2024-08-19T08:23:09Z) - Attention Based Simple Primitives for Open World Compositional Zero-Shot Learning [12.558701595138928]
Compositional Zero-Shot Learning (CZSL) aims to predict unknown compositions made up of attribute and object pairs.
We are exploring Open World Compositional Zero-Shot Learning (OW-CZSL) in this study, where our test space encompasses all potential combinations of attributes and objects.
Our approach involves utilizing the self-attention mechanism between attributes and objects to achieve better generalization from seen to unseen compositions.
arXiv Detail & Related papers (2024-07-18T17:11:29Z) - ProCC: Progressive Cross-primitive Compatibility for Open-World
Compositional Zero-Shot Learning [29.591615811894265]
Open-World Compositional Zero-shot Learning (OW-CZSL) aims to recognize novel compositions of state and object primitives in images with no priors on the compositional space.
We propose a novel method, termed Progressive Cross-primitive Compatibility (ProCC), to mimic the human learning process for OW-CZSL tasks.
arXiv Detail & Related papers (2022-11-19T10:09:46Z) - Simple Primitives with Feasibility- and Contextuality-Dependence for
Open-World Compositional Zero-shot Learning [86.5258816031722]
The task of Compositional Zero-Shot Learning (CZSL) is to recognize images of novel state-object compositions that are absent during the training stage.
Previous methods of learning compositional embedding have shown effectiveness in closed-world CZSL.
In Open-World CZSL (OW-CZSL), their performance tends to degrade significantly due to the large cardinality of possible compositions.
arXiv Detail & Related papers (2022-11-05T12:57:06Z) - Reference-Limited Compositional Zero-Shot Learning [19.10692212692771]
Compositional zero-shot learning (CZSL) refers to recognizing unseen compositions of known visual primitives.
We propose a novel Meta Compositional Graph Learner (MetaCGL) that can efficiently learn the compositionality from insufficient referential information.
arXiv Detail & Related papers (2022-08-22T03:58:02Z) - Siamese Contrastive Embedding Network for Compositional Zero-Shot
Learning [76.13542095170911]
Compositional Zero-Shot Learning (CZSL) aims to recognize unseen compositions formed from seen state and object during training.
We propose a novel Siamese Contrastive Embedding Network (SCEN) for unseen composition recognition.
Our method significantly outperforms the state-of-the-art approaches on three challenging benchmark datasets.
arXiv Detail & Related papers (2022-06-29T09:02:35Z) - KG-SP: Knowledge Guided Simple Primitives for Open World Compositional
Zero-Shot Learning [52.422873819371276]
The goal of open-world compositional zero-shot learning (OW-CZSL) is to recognize compositions of state and objects in images.
Here, we revisit a simple CZSL baseline and predict the primitives, i.e. states and objects, independently.
We estimate the feasibility of each composition through external knowledge, using this prior to remove unfeasible compositions from the output space.
Our model, Knowledge-Guided Simple Primitives (KG-SP), achieves state of the art in both OW-CZSL and pCZSL.
arXiv Detail & Related papers (2022-05-13T17:18:15Z) - Learning Graph Embeddings for Open World Compositional Zero-Shot
Learning [47.09665742252187]
Compositional Zero-Shot learning (CZSL) aims to recognize unseen compositions of state and object visual primitives seen during training.
We propose a new approach, Compositional Cosine Graph Embeddings (Co-CGE)
Co-CGE models the dependency between states, objects and their compositions through a graph convolutional neural network.
arXiv Detail & Related papers (2021-05-03T17:08:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.