Reference-Limited Compositional Zero-Shot Learning
- URL: http://arxiv.org/abs/2208.10046v2
- Date: Sat, 29 Apr 2023 14:10:32 GMT
- Title: Reference-Limited Compositional Zero-Shot Learning
- Authors: Siteng Huang, Qiyao Wei, Donglin Wang
- Abstract summary: Compositional zero-shot learning (CZSL) refers to recognizing unseen compositions of known visual primitives.
We propose a novel Meta Compositional Graph Learner (MetaCGL) that can efficiently learn the compositionality from insufficient referential information.
- Score: 19.10692212692771
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Compositional zero-shot learning (CZSL) refers to recognizing unseen
compositions of known visual primitives, which is an essential ability for
artificial intelligence systems to learn and understand the world. While
considerable progress has been made on existing benchmarks, we suspect whether
popular CZSL methods can address the challenges of few-shot and few referential
compositions, which is common when learning in real-world unseen environments.
To this end, we study the challenging reference-limited compositional zero-shot
learning (RL-CZSL) problem in this paper, i.e., given limited seen compositions
that contain only a few samples as reference, unseen compositions of observed
primitives should be identified. We propose a novel Meta Compositional Graph
Learner (MetaCGL) that can efficiently learn the compositionality from
insufficient referential information and generalize to unseen compositions.
Besides, we build a benchmark with two new large-scale datasets that consist of
natural images with diverse compositional labels, providing more realistic
environments for RL-CZSL. Extensive experiments in the benchmarks show that our
method achieves state-of-the-art performance in recognizing unseen compositions
when reference is limited for compositional learning.
Related papers
- Cross-composition Feature Disentanglement for Compositional Zero-shot Learning [49.919635694894204]
Disentanglement of visual features of primitives (i.e., attributes and objects) has shown exceptional results in Compositional Zero-shot Learning (CZSL)
We propose the solution of cross-composition feature disentanglement, which takes multiple primitive-sharing compositions as inputs and constrains the disentangled primitive features to be general across these compositions.
arXiv Detail & Related papers (2024-08-19T08:23:09Z) - Mutual Balancing in State-Object Components for Compositional Zero-Shot
Learning [0.0]
Compositional Zero-Shot Learning (CZSL) aims to recognize unseen compositions from seen states and objects.
We propose a novel method called MUtual balancing in STate-object components (MUST) for CZSL, which provides a balancing inductive bias for the model.
Our approach significantly outperforms the state-of-the-art on MIT-States, UT-Zappos, and C-GQA when combined with the basic CZSL frameworks.
arXiv Detail & Related papers (2022-11-19T10:21:22Z) - Learning Attention Propagation for Compositional Zero-Shot Learning [71.55375561183523]
We propose a novel method called Compositional Attention Propagated Embedding (CAPE)
CAPE learns to identify this structure and propagates knowledge between them to learn class embedding for all seen and unseen compositions.
We show that our method outperforms previous baselines to set a new state-of-the-art on three publicly available benchmarks.
arXiv Detail & Related papers (2022-10-20T19:44:11Z) - KG-SP: Knowledge Guided Simple Primitives for Open World Compositional
Zero-Shot Learning [52.422873819371276]
The goal of open-world compositional zero-shot learning (OW-CZSL) is to recognize compositions of state and objects in images.
Here, we revisit a simple CZSL baseline and predict the primitives, i.e. states and objects, independently.
We estimate the feasibility of each composition through external knowledge, using this prior to remove unfeasible compositions from the output space.
Our model, Knowledge-Guided Simple Primitives (KG-SP), achieves state of the art in both OW-CZSL and pCZSL.
arXiv Detail & Related papers (2022-05-13T17:18:15Z) - Learning Graph Embeddings for Open World Compositional Zero-Shot
Learning [47.09665742252187]
Compositional Zero-Shot learning (CZSL) aims to recognize unseen compositions of state and object visual primitives seen during training.
We propose a new approach, Compositional Cosine Graph Embeddings (Co-CGE)
Co-CGE models the dependency between states, objects and their compositions through a graph convolutional neural network.
arXiv Detail & Related papers (2021-05-03T17:08:21Z) - Learning Graph Embeddings for Compositional Zero-shot Learning [73.80007492964951]
In compositional zero-shot learning, the goal is to recognize unseen compositions of observed visual primitives states.
We propose a novel graph formulation called Compositional Graph Embedding (CGE) that learns image features and latent representations of visual primitives in an end-to-end manner.
By learning a joint compatibility that encodes semantics between concepts, our model allows for generalization to unseen compositions without relying on an external knowledge base like WordNet.
arXiv Detail & Related papers (2021-02-03T10:11:03Z) - Open World Compositional Zero-Shot Learning [47.09665742252187]
Compositional Zero-Shot learning (CZSL) requires to recognize state-object compositions unseen during training.
We operate on the open world setting, where the search space includes a large number of unseen compositions.
arXiv Detail & Related papers (2021-01-29T14:45:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.