CURI: A Benchmark for Productive Concept Learning Under Uncertainty
- URL: http://arxiv.org/abs/2010.02855v1
- Date: Tue, 6 Oct 2020 16:23:17 GMT
- Title: CURI: A Benchmark for Productive Concept Learning Under Uncertainty
- Authors: Ramakrishna Vedantam, Arthur Szlam, Maximilian Nickel, Ari Morcos,
Brenden Lake
- Abstract summary: We introduce a new few-shot, meta-learning benchmark, Compositional Reasoning Under Uncertainty (CURI)
CURI evaluates different aspects of productive and systematic generalization, including abstract understandings of disentangling, productive generalization, learning operations, variable binding, etc.
It also defines a model-independent "compositionality gap" to evaluate the difficulty of generalizing out-of-distribution along each of these axes.
- Score: 33.83721664338612
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Humans can learn and reason under substantial uncertainty in a space of
infinitely many concepts, including structured relational concepts ("a scene
with objects that have the same color") and ad-hoc categories defined through
goals ("objects that could fall on one's head"). In contrast, standard
classification benchmarks: 1) consider only a fixed set of category labels, 2)
do not evaluate compositional concept learning and 3) do not explicitly capture
a notion of reasoning under uncertainty. We introduce a new few-shot,
meta-learning benchmark, Compositional Reasoning Under Uncertainty (CURI) to
bridge this gap. CURI evaluates different aspects of productive and systematic
generalization, including abstract understandings of disentangling, productive
generalization, learning boolean operations, variable binding, etc.
Importantly, it also defines a model-independent "compositionality gap" to
evaluate the difficulty of generalizing out-of-distribution along each of these
axes. Extensive evaluations across a range of modeling choices spanning
different modalities (image, schemas, and sounds), splits, privileged auxiliary
concept information, and choices of negatives reveal substantial scope for
modeling advances on the proposed task. All code and datasets will be available
online.
Related papers
- Evaluating Readability and Faithfulness of Concept-based Explanations [35.48852504832633]
Concept-based explanations arise as a promising avenue for explaining high-level patterns learned by Large Language Models.
Current methods approach concepts from different perspectives, lacking a unified formalization.
This makes evaluating the core measures of concepts, namely faithfulness or readability, challenging.
arXiv Detail & Related papers (2024-04-29T09:20:25Z) - Generalized Unbiased Scene Graph Generation [85.22334551067617]
Generalized Unbiased Scene Graph Generation (G-USGG) takes into account both predicate-level and concept-level imbalance.
We propose the Multi-Concept Learning (MCL) framework, which ensures a balanced learning process across rare/ uncommon/ common concepts.
arXiv Detail & Related papers (2023-08-09T08:51:03Z) - Concept-Based Explanations to Test for False Causal Relationships
Learned by Abusive Language Classifiers [7.022948483613113]
We consider three well-known abusive language classifiers trained on large English datasets.
We first examine the unwanted dependencies learned by the classifiers by assessing their accuracy on a challenge set across all decision thresholds.
We then introduce concept-based explanation metrics to assess the influence of the concept on the labels.
arXiv Detail & Related papers (2023-07-04T19:57:54Z) - Divide and Conquer: Answering Questions with Object Factorization and
Compositional Reasoning [30.392986232906107]
We propose an integral framework consisting of a principled object factorization method and a novel neural module network.
Our factorization method decomposes objects based on their key characteristics, and automatically derives prototypes that represent a wide range of objects.
With these prototypes encoding important semantics, the proposed network then correlates objects by measuring their similarity on a common semantic space.
It is capable of answering questions with diverse objects regardless of their availability during training, and overcoming the issues of biased question-answer distributions.
arXiv Detail & Related papers (2023-03-18T19:37:28Z) - Succinct Representations for Concepts [12.134564449202708]
Foundation models like chatGPT have demonstrated remarkable performance on various tasks.
However, for many questions, they may produce false answers that look accurate.
In this paper, we introduce succinct representations of concepts based on category theory.
arXiv Detail & Related papers (2023-03-01T12:11:23Z) - Learning What Not to Segment: A New Perspective on Few-Shot Segmentation [63.910211095033596]
Recently few-shot segmentation (FSS) has been extensively developed.
This paper proposes a fresh and straightforward insight to alleviate the problem.
In light of the unique nature of the proposed approach, we also extend it to a more realistic but challenging setting.
arXiv Detail & Related papers (2022-03-15T03:08:27Z) - Resolving label uncertainty with implicit posterior models [71.62113762278963]
We propose a method for jointly inferring labels across a collection of data samples.
By implicitly assuming the existence of a generative model for which a differentiable predictor is the posterior, we derive a training objective that allows learning under weak beliefs.
arXiv Detail & Related papers (2022-02-28T18:09:44Z) - Translational Concept Embedding for Generalized Compositional Zero-shot
Learning [73.60639796305415]
Generalized compositional zero-shot learning means to learn composed concepts of attribute-object pairs in a zero-shot fashion.
This paper introduces a new approach, termed translational concept embedding, to solve these two difficulties in a unified framework.
arXiv Detail & Related papers (2021-12-20T21:27:51Z) - A Minimalist Dataset for Systematic Generalization of Perception,
Syntax, and Semantics [131.93113552146195]
We present a new dataset, Handwritten arithmetic with INTegers (HINT), to examine machines' capability of learning generalizable concepts.
In HINT, machines are tasked with learning how concepts are perceived from raw signals such as images.
We undertake extensive experiments with various sequence-to-sequence models, including RNNs, Transformers, and GPT-3.
arXiv Detail & Related papers (2021-03-02T01:32:54Z) - Concept Learners for Few-Shot Learning [76.08585517480807]
We propose COMET, a meta-learning method that improves generalization ability by learning to learn along human-interpretable concept dimensions.
We evaluate our model on few-shot tasks from diverse domains, including fine-grained image classification, document categorization and cell type annotation.
arXiv Detail & Related papers (2020-07-14T22:04:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.