ZeroC: A Neuro-Symbolic Model for Zero-shot Concept Recognition and
Acquisition at Inference Time
- URL: http://arxiv.org/abs/2206.15049v2
- Date: Sun, 3 Jul 2022 01:48:44 GMT
- Title: ZeroC: A Neuro-Symbolic Model for Zero-shot Concept Recognition and
Acquisition at Inference Time
- Authors: Tailin Wu, Megan Tjandrasuwita, Zhengxuan Wu, Xuelin Yang, Kevin Liu,
Rok Sosi\v{c}, Jure Leskovec
- Abstract summary: Humans have the remarkable ability to recognize and acquire novel visual concepts in a zero-shot manner.
We introduce Zero-shot Concept Recognition and Acquisition (ZeroC), a neuro-symbolic architecture that can recognize and acquire novel concepts in a zero-shot way.
- Score: 49.067846763204564
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Humans have the remarkable ability to recognize and acquire novel visual
concepts in a zero-shot manner. Given a high-level, symbolic description of a
novel concept in terms of previously learned visual concepts and their
relations, humans can recognize novel concepts without seeing any examples.
Moreover, they can acquire new concepts by parsing and communicating symbolic
structures using learned visual concepts and relations. Endowing these
capabilities in machines is pivotal in improving their generalization
capability at inference time. In this work, we introduce Zero-shot Concept
Recognition and Acquisition (ZeroC), a neuro-symbolic architecture that can
recognize and acquire novel concepts in a zero-shot way. ZeroC represents
concepts as graphs of constituent concept models (as nodes) and their relations
(as edges). To allow inference time composition, we employ energy-based models
(EBMs) to model concepts and relations. We design ZeroC architecture so that it
allows a one-to-one mapping between a symbolic graph structure of a concept and
its corresponding EBM, which for the first time, allows acquiring new concepts,
communicating its graph structure, and applying it to classification and
detection tasks (even across domains) at inference time. We introduce
algorithms for learning and inference with ZeroC. We evaluate ZeroC on a
challenging grid-world dataset which is designed to probe zero-shot concept
recognition and acquisition, and demonstrate its capability.
Related papers
- Exploiting Interpretable Capabilities with Concept-Enhanced Diffusion and Prototype Networks [8.391254800873599]
We create concept-enriched models that incorporate concept information into existing architectures.
In particular, we propose Concept-Guided Diffusion Conditional, which can generate visual representations of concepts, and Concept-Guided Prototype Networks, which can create a concept prototype dataset and leverage it to perform interpretable concept prediction.
These results open up new lines of research by exploiting pre-existing information in the quest for rendering machine learning more human-understandable.
arXiv Detail & Related papers (2024-10-24T13:07:56Z) - Neural Concept Binder [22.074896812195437]
We introduce the Neural Concept Binder (NCB), a framework for deriving both discrete and continuous concept representations.
The structured nature of NCB's concept representations allows for intuitive inspection and the straightforward integration of external knowledge.
We validate the effectiveness of NCB through evaluations on our newly introduced CLEVR-Sudoku dataset.
arXiv Detail & Related papers (2024-06-14T11:52:09Z) - LLM-assisted Concept Discovery: Automatically Identifying and Explaining Neuron Functions [15.381209058506078]
Prior works have associated concepts with neurons based on examples of concepts or a pre-defined set of concepts.
We propose to leverage multimodal large language models for automatic and open-ended concept discovery.
We validate each concept by generating examples and counterexamples and evaluating the neuron's response on this new set of images.
arXiv Detail & Related papers (2024-06-12T18:19:37Z) - Knowledge graphs for empirical concept retrieval [1.06378109904813]
Concept-based explainable AI is promising as a tool to improve the understanding of complex models at the premises of a given user.
Here, we present a workflow for user-driven data collection in both text and image domains.
We test the retrieved concept datasets on two concept-based explainability methods, namely concept activation vectors (CAVs) and concept activation regions (CARs)
arXiv Detail & Related papers (2024-04-10T13:47:22Z) - CEIR: Concept-based Explainable Image Representation Learning [0.4198865250277024]
We introduce Concept-based Explainable Image Representation (CEIR) to derive high-quality representations without label dependency.
Our method exhibits state-of-the-art unsupervised clustering performance on benchmarks such as CIFAR10, CIFAR100, and STL10.
CEIR can seamlessly extract the related concept from open-world images without fine-tuning.
arXiv Detail & Related papers (2023-12-17T15:37:41Z) - A Recursive Bateson-Inspired Model for the Generation of Semantic Formal
Concepts from Spatial Sensory Data [77.34726150561087]
This paper presents a new symbolic-only method for the generation of hierarchical concept structures from complex sensory data.
The approach is based on Bateson's notion of difference as the key to the genesis of an idea or a concept.
The model is able to produce fairly rich yet human-readable conceptual representations without training.
arXiv Detail & Related papers (2023-07-16T15:59:13Z) - ConceptBed: Evaluating Concept Learning Abilities of Text-to-Image
Diffusion Models [79.10890337599166]
We introduce ConceptBed, a large-scale dataset that consists of 284 unique visual concepts and 33K composite text prompts.
We evaluate visual concepts that are either objects, attributes, or styles, and also evaluate four dimensions of compositionality: counting, attributes, relations, and actions.
Our results point to a trade-off between learning the concepts and preserving the compositionality which existing approaches struggle to overcome.
arXiv Detail & Related papers (2023-06-07T18:00:38Z) - Human-Centered Concept Explanations for Neural Networks [47.71169918421306]
We introduce concept explanations including the class of Concept Activation Vectors (CAV)
We then discuss approaches to automatically extract concepts, and approaches to address some of their caveats.
Finally, we discuss some case studies that showcase the utility of such concept-based explanations in synthetic settings and real world applications.
arXiv Detail & Related papers (2022-02-25T01:27:31Z) - Formalising Concepts as Grounded Abstractions [68.24080871981869]
This report shows how representation learning can be used to induce concepts from raw data.
The main technical goal of this report is to show how techniques from representation learning can be married with a lattice-theoretic formulation of conceptual spaces.
arXiv Detail & Related papers (2021-01-13T15:22:01Z) - Interpretable Visual Reasoning via Induced Symbolic Space [75.95241948390472]
We study the problem of concept induction in visual reasoning, i.e., identifying concepts and their hierarchical relationships from question-answer pairs associated with images.
We first design a new framework named object-centric compositional attention model (OCCAM) to perform the visual reasoning task with object-level visual features.
We then come up with a method to induce concepts of objects and relations using clues from the attention patterns between objects' visual features and question words.
arXiv Detail & Related papers (2020-11-23T18:21:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.