Learning Algebraic Recombination for Compositional Generalization
- URL: http://arxiv.org/abs/2107.06516v1
- Date: Wed, 14 Jul 2021 07:23:46 GMT
- Title: Learning Algebraic Recombination for Compositional Generalization
- Authors: Chenyao Liu, Shengnan An, Zeqi Lin, Qian Liu, Bei Chen, Jian-Guang
Lou, Lijie Wen, Nanning Zheng and Dongmei Zhang
- Abstract summary: We propose LeAR, an end-to-end neural model to learn algebraic recombination for compositional generalization.
Key insight is to model the semantic parsing task as a homomorphism between a latent syntactic algebra and a semantic algebra.
Experiments on two realistic and comprehensive compositional generalization demonstrate the effectiveness of our model.
- Score: 71.78771157219428
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural sequence models exhibit limited compositional generalization ability
in semantic parsing tasks. Compositional generalization requires algebraic
recombination, i.e., dynamically recombining structured expressions in a
recursive manner. However, most previous studies mainly concentrate on
recombining lexical units, which is an important but not sufficient part of
algebraic recombination. In this paper, we propose LeAR, an end-to-end neural
model to learn algebraic recombination for compositional generalization. The
key insight is to model the semantic parsing task as a homomorphism between a
latent syntactic algebra and a semantic algebra, thus encouraging algebraic
recombination. Specifically, we learn two modules jointly: a Composer for
producing latent syntax, and an Interpreter for assigning semantic operations.
Experiments on two realistic and comprehensive compositional generalization
benchmarks demonstrate the effectiveness of our model. The source code is
publicly available at https://github.com/microsoft/ContextualSP.
Related papers
- Compositional Generalization Requires Compositional Parsers [69.77216620997305]
We compare sequence-to-sequence models and models guided by compositional principles on the recent COGS corpus.
We show structural generalization is a key measure of compositional generalization and requires models that are aware of complex structure.
arXiv Detail & Related papers (2022-02-24T07:36:35Z) - Learning Algebraic Representation for Systematic Generalization in
Abstract Reasoning [109.21780441933164]
We propose a hybrid approach to improve systematic generalization in reasoning.
We showcase a prototype with algebraic representation for the abstract spatial-temporal task of Raven's Progressive Matrices (RPM)
We show that the algebraic representation learned can be decoded by isomorphism to generate an answer.
arXiv Detail & Related papers (2021-11-25T09:56:30Z) - Grounded Graph Decoding Improves Compositional Generalization in
Question Answering [68.72605660152101]
Question answering models struggle to generalize to novel compositions of training patterns, such as longer sequences or more complex test structures.
We propose Grounded Graph Decoding, a method to improve compositional generalization of language representations by grounding structured predictions with an attention mechanism.
Our model significantly outperforms state-of-the-art baselines on the Compositional Freebase Questions (CFQ) dataset, a challenging benchmark for compositional generalization in question answering.
arXiv Detail & Related papers (2021-11-05T17:50:14Z) - Improving Compositional Generalization in Classification Tasks via
Structure Annotations [33.90268697120572]
Humans have a great ability to generalize compositionally, but state-of-the-art neural models struggle to do so.
First, we study ways to convert a natural language sequence-to-sequence dataset to a classification dataset that also requires compositional generalization.
Second, we show that providing structural hints (specifically, providing parse trees and entity links as attention masks for a Transformer model) helps compositional generalization.
arXiv Detail & Related papers (2021-06-19T06:07:27Z) - Abelian Neural Networks [48.52497085313911]
We first construct a neural network architecture for Abelian group operations and derive a universal approximation property.
We extend it to Abelian semigroup operations using the characterization of associative symmetrics.
We train our models over fixed word embeddings and demonstrate improved performance over the original word2vec.
arXiv Detail & Related papers (2021-02-24T11:52:21Z) - General Data Analytics with Applications to Visual Information Analysis:
A Provable Backward-Compatible Semisimple Paradigm over T-Algebra [7.028302194243312]
We consider a novel backward-compatible paradigm of general data analytics over a recently-reported semisimple algebra.
We generalize some canonical algorithms for visual pattern analysis.
Experiments on public datasets show that the generalized algorithms compare favorably with their canonical counterparts.
arXiv Detail & Related papers (2020-10-31T16:41:09Z) - Compositional Generalization by Learning Analytical Expressions [87.15737632096378]
A memory-augmented neural model is connected with analytical expressions to achieve compositional generalization.
Experiments on the well-known benchmark SCAN demonstrate that our model seizes a great ability of compositional generalization.
arXiv Detail & Related papers (2020-06-18T15:50:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.