Concepts, Properties and an Approach for Compositional Generalization
- URL: http://arxiv.org/abs/2102.04225v1
- Date: Mon, 8 Feb 2021 14:22:30 GMT
- Title: Concepts, Properties and an Approach for Compositional Generalization
- Authors: Yuanpeng Li
- Abstract summary: This report connects a series of our work for compositional generalization, and summarizes an approach.
The approach uses architecture design and regularization to regulate information of representations.
We hope this work would be helpful to clarify fundamentals of compositional generalization and lead to advance artificial intelligence.
- Score: 2.0559497209595823
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Compositional generalization is the capacity to recognize and imagine a large
amount of novel combinations from known components. It is a key in human
intelligence, but current neural networks generally lack such ability. This
report connects a series of our work for compositional generalization, and
summarizes an approach. The first part contains concepts and properties. The
second part looks into a machine learning approach. The approach uses
architecture design and regularization to regulate information of
representations. This report focuses on basic ideas with intuitive and
illustrative explanations. We hope this work would be helpful to clarify
fundamentals of compositional generalization and lead to advance artificial
intelligence.
Related papers
- A Complexity-Based Theory of Compositionality [53.025566128892066]
In AI, compositional representations can enable a powerful form of out-of-distribution generalization.
Here, we propose a formal definition of compositionality that accounts for and extends our intuitions about compositionality.
The definition is conceptually simple, quantitative, grounded in algorithmic information theory, and applicable to any representation.
arXiv Detail & Related papers (2024-10-18T18:37:27Z) - Improving Compositional Generalization Using Iterated Learning and
Simplicial Embeddings [19.667133565610087]
Compositional generalization is easy for humans but hard for deep neural networks.
We propose to improve this ability by using iterated learning on models with simplicial embeddings.
We show that this combination of changes improves compositional generalization over other approaches.
arXiv Detail & Related papers (2023-10-28T18:30:30Z) - Provable Compositional Generalization for Object-Centric Learning [57.42720932595342]
Learning representations that generalize to novel compositions of known concepts is crucial for bridging the gap between human and machine perception.
We show that autoencoders that satisfy structural assumptions on the decoder and enforce encoder-decoder consistency will learn object-centric representations that provably generalize compositionally.
arXiv Detail & Related papers (2023-10-09T01:18:07Z) - Vector-based Representation is the Key: A Study on Disentanglement and
Compositional Generalization [77.57425909520167]
We show that it is possible to achieve both good concept recognition and novel concept composition.
We propose a method to reform the scalar-based disentanglement works to be vector-based to increase both capabilities.
arXiv Detail & Related papers (2023-05-29T13:05:15Z) - Separating Skills and Concepts for Novel Visual Question Answering [66.46070380927372]
Generalization to out-of-distribution data has been a problem for Visual Question Answering (VQA) models.
"Skills" are visual tasks, such as counting or attribute recognition, and are applied to "concepts" mentioned in the question.
We present a novel method for learning to compose skills and concepts that separates these two factors implicitly within a model.
arXiv Detail & Related papers (2021-07-19T18:55:10Z) - Improving Compositional Generalization in Classification Tasks via
Structure Annotations [33.90268697120572]
Humans have a great ability to generalize compositionally, but state-of-the-art neural models struggle to do so.
First, we study ways to convert a natural language sequence-to-sequence dataset to a classification dataset that also requires compositional generalization.
Second, we show that providing structural hints (specifically, providing parse trees and entity links as attention masks for a Transformer model) helps compositional generalization.
arXiv Detail & Related papers (2021-06-19T06:07:27Z) - Flexible Compositional Learning of Structured Visual Concepts [17.665938343060112]
We study how people learn different types of visual compositions, using abstract visual forms with rich relational structure.
We find that people can make meaningful compositional generalizations from just a few examples in a variety of scenarios.
Unlike past work examining special cases of compositionality, our work shows how a single computational approach can account for many distinct types of compositional generalization.
arXiv Detail & Related papers (2021-05-20T15:48:05Z) - Compositional Processing Emerges in Neural Networks Solving Math
Problems [100.80518350845668]
Recent progress in artificial neural networks has shown that when large models are trained on enough linguistic data, grammatical structure emerges in their representations.
We extend this work to the domain of mathematical reasoning, where it is possible to formulate precise hypotheses about how meanings should be composed.
Our work shows that neural networks are not only able to infer something about the structured relationships implicit in their training data, but can also deploy this knowledge to guide the composition of individual meanings into composite wholes.
arXiv Detail & Related papers (2021-05-19T07:24:42Z) - Compositional Generalization by Learning Analytical Expressions [87.15737632096378]
A memory-augmented neural model is connected with analytical expressions to achieve compositional generalization.
Experiments on the well-known benchmark SCAN demonstrate that our model seizes a great ability of compositional generalization.
arXiv Detail & Related papers (2020-06-18T15:50:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.