Translational Concept Embedding for Generalized Compositional Zero-shot
Learning
- URL: http://arxiv.org/abs/2112.10871v1
- Date: Mon, 20 Dec 2021 21:27:51 GMT
- Title: Translational Concept Embedding for Generalized Compositional Zero-shot
Learning
- Authors: He Huang, Wei Tang, Jiawei Zhang, Philip S. Yu
- Abstract summary: Generalized compositional zero-shot learning means to learn composed concepts of attribute-object pairs in a zero-shot fashion.
This paper introduces a new approach, termed translational concept embedding, to solve these two difficulties in a unified framework.
- Score: 73.60639796305415
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generalized compositional zero-shot learning means to learn composed concepts
of attribute-object pairs in a zero-shot fashion, where a model is trained on a
set of seen concepts and tested on a combined set of seen and unseen concepts.
This task is very challenging because of not only the gap between seen and
unseen concepts but also the contextual dependency between attributes and
objects. This paper introduces a new approach, termed translational concept
embedding, to solve these two difficulties in a unified framework. It models
the effect of applying an attribute to an object as adding a translational
attribute feature to an object prototype. We explicitly take into account of
the contextual dependency between attributes and objects by generating
translational attribute features conditionally dependent on the object
prototypes. Furthermore, we design a ratio variance constraint loss to promote
the model's generalization ability on unseen concepts. It regularizes the
distances between concepts by utilizing knowledge from their pretrained word
embeddings. We evaluate the performance of our model under both the unbiased
and biased concept classification tasks, and show that our model is able to
achieve good balance in predicting unseen and seen concepts.
Related papers
- LLM-based Hierarchical Concept Decomposition for Interpretable Fine-Grained Image Classification [5.8754760054410955]
We introduce textttHi-CoDecomposition, a novel framework designed to enhance model interpretability through structured concept analysis.
Our approach not only aligns with the performance of state-of-the-art models but also advances transparency by providing clear insights into the decision-making process.
arXiv Detail & Related papers (2024-05-29T00:36:56Z) - Advancing Ante-Hoc Explainable Models through Generative Adversarial Networks [24.45212348373868]
This paper presents a novel concept learning framework for enhancing model interpretability and performance in visual classification tasks.
Our approach appends an unsupervised explanation generator to the primary classifier network and makes use of adversarial training.
This work presents a significant step towards building inherently interpretable deep vision models with task-aligned concept representations.
arXiv Detail & Related papers (2024-01-09T16:16:16Z) - Interpreting Pretrained Language Models via Concept Bottlenecks [55.47515772358389]
Pretrained language models (PLMs) have made significant strides in various natural language processing tasks.
The lack of interpretability due to their black-box'' nature poses challenges for responsible implementation.
We propose a novel approach to interpreting PLMs by employing high-level, meaningful concepts that are easily understandable for humans.
arXiv Detail & Related papers (2023-11-08T20:41:18Z) - Concept-Centric Transformers: Enhancing Model Interpretability through
Object-Centric Concept Learning within a Shared Global Workspace [1.6574413179773757]
Concept-Centric Transformers is a simple yet effective configuration of the shared global workspace for interpretability.
We show that our model achieves better classification accuracy than all baselines across all problems.
arXiv Detail & Related papers (2023-05-25T06:37:39Z) - Context-LGM: Leveraging Object-Context Relation for Context-Aware Object
Recognition [48.5398871460388]
We propose a novel Contextual Latent Generative Model (Context-LGM), which considers the object-context relation and models it in a hierarchical manner.
To infer contextual features, we reformulate the objective function of Variational Auto-Encoder (VAE), where contextual features are learned as a posterior conditioned distribution on the object.
The effectiveness of our method is verified by state-of-the-art performance on two context-aware object recognition tasks.
arXiv Detail & Related papers (2021-10-08T11:31:58Z) - Relation-aware Compositional Zero-shot Learning for Attribute-Object
Pair Recognition [17.464548471883948]
This paper proposes a novel model for recognizing images with composite attribute-object concepts.
We aim to explore the three key properties required to learn rich and robust features for primitive concepts that compose attribute-object pairs.
To prevent the model from being biased towards seen composite concepts and reduce the entanglement between attributes and objects, we propose a blocking mechanism.
arXiv Detail & Related papers (2021-08-10T11:23:03Z) - Separating Skills and Concepts for Novel Visual Question Answering [66.46070380927372]
Generalization to out-of-distribution data has been a problem for Visual Question Answering (VQA) models.
"Skills" are visual tasks, such as counting or attribute recognition, and are applied to "concepts" mentioned in the question.
We present a novel method for learning to compose skills and concepts that separates these two factors implicitly within a model.
arXiv Detail & Related papers (2021-07-19T18:55:10Z) - Understanding Synonymous Referring Expressions via Contrastive Features [105.36814858748285]
We develop an end-to-end trainable framework to learn contrastive features on the image and object instance levels.
We conduct extensive experiments to evaluate the proposed algorithm on several benchmark datasets.
arXiv Detail & Related papers (2021-04-20T17:56:24Z) - Concept Learners for Few-Shot Learning [76.08585517480807]
We propose COMET, a meta-learning method that improves generalization ability by learning to learn along human-interpretable concept dimensions.
We evaluate our model on few-shot tasks from diverse domains, including fine-grained image classification, document categorization and cell type annotation.
arXiv Detail & Related papers (2020-07-14T22:04:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.