Iterated learning for emergent systematicity in VQA
- URL: http://arxiv.org/abs/2105.01119v1
- Date: Mon, 3 May 2021 18:44:06 GMT
- Title: Iterated learning for emergent systematicity in VQA
- Authors: Ankit Vani, Max Schwarzer, Yuchen Lu, Eeshan Dhekane, Aaron Courville
- Abstract summary: neural module networks have an architectural bias towards compositionality.
When learning layouts and modules jointly, compositionality does not arise automatically and an explicit pressure is necessary for the emergence of layouts exhibiting the right structure.
We propose to address this problem using iterated learning, a cognitive science theory of the emergence of compositional languages in nature.
- Score: 3.977144385787228
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although neural module networks have an architectural bias towards
compositionality, they require gold standard layouts to generalize
systematically in practice. When instead learning layouts and modules jointly,
compositionality does not arise automatically and an explicit pressure is
necessary for the emergence of layouts exhibiting the right structure. We
propose to address this problem using iterated learning, a cognitive science
theory of the emergence of compositional languages in nature that has primarily
been applied to simple referential games in machine learning. Considering the
layouts of module networks as samples from an emergent language, we use
iterated learning to encourage the development of structure within this
language. We show that the resulting layouts support systematic generalization
in neural agents solving the more complex task of visual question-answering.
Our regularized iterated learning method can outperform baselines without
iterated learning on SHAPES-SyGeT (SHAPES Systematic Generalization Test), a
new split of the SHAPES dataset we introduce to evaluate systematic
generalization, and on CLOSURE, an extension of CLEVR also designed to test
systematic generalization. We demonstrate superior performance in recovering
ground-truth compositional program structure with limited supervision on both
SHAPES-SyGeT and CLEVR.
Related papers
- On The Specialization of Neural Modules [16.83151955540625]
We study the ability of network modules to specialize to useful structures in a dataset and achieve systematic generalization.
Our results shed light on the difficulty of module specialization, what is required for modules to successfully specialize, and the necessity of modular architectures to achieve systematicity.
arXiv Detail & Related papers (2024-09-23T12:58:11Z) - On the Role of Information Structure in Reinforcement Learning for Partially-Observable Sequential Teams and Games [55.2480439325792]
In a sequential decision-making problem, the information structure is the description of how events in the system occurring at different points in time affect each other.
By contrast, real-world sequential decision-making problems typically involve a complex and time-varying interdependence of system variables.
We formalize a novel reinforcement learning model which explicitly represents the information structure.
arXiv Detail & Related papers (2024-03-01T21:28:19Z) - Inducing Systematicity in Transformers by Attending to Structurally
Quantized Embeddings [60.698130703909804]
Transformers generalize to novel compositions of structures and entities after being trained on a complex dataset.
We propose SQ-Transformer that explicitly encourages systematicity in the embeddings and attention layers.
We show that SQ-Transformer achieves stronger compositional generalization than the vanilla Transformer on multiple low-complexity semantic parsing and machine translation datasets.
arXiv Detail & Related papers (2024-02-09T15:53:15Z) - Neural-Symbolic Recursive Machine for Systematic Generalization [113.22455566135757]
We introduce the Neural-Symbolic Recursive Machine (NSR), whose core is a Grounded Symbol System (GSS)
NSR integrates neural perception, syntactic parsing, and semantic reasoning.
We evaluate NSR's efficacy across four challenging benchmarks designed to probe systematic generalization capabilities.
arXiv Detail & Related papers (2022-10-04T13:27:38Z) - Learning Dynamics and Structure of Complex Systems Using Graph Neural
Networks [13.509027957413409]
We trained graph neural networks to fit time series from an example nonlinear dynamical system.
We found simple interpretations of the learned representation and model components.
We successfully identified a graph translator' between the statistical interactions in belief propagation and parameters of the corresponding trained network.
arXiv Detail & Related papers (2022-02-22T15:58:16Z) - Provable Hierarchy-Based Meta-Reinforcement Learning [50.17896588738377]
We analyze HRL in the meta-RL setting, where learner learns latent hierarchical structure during meta-training for use in a downstream task.
We provide "diversity conditions" which, together with a tractable optimism-based algorithm, guarantee sample-efficient recovery of this natural hierarchy.
Our bounds incorporate common notions in HRL literature such as temporal and state/action abstractions, suggesting that our setting and analysis capture important features of HRL in practice.
arXiv Detail & Related papers (2021-10-18T17:56:02Z) - Structure-Aware Feature Generation for Zero-Shot Learning [108.76968151682621]
We introduce a novel structure-aware feature generation scheme, termed as SA-GAN, to account for the topological structure in learning both the latent space and the generative networks.
Our method significantly enhances the generalization capability on unseen-classes and consequently improve the classification performance.
arXiv Detail & Related papers (2021-08-16T11:52:08Z) - Systematic Generalization on gSCAN with Language Conditioned Embedding [19.39687991647301]
Systematic Generalization refers to a learning algorithm's ability to extrapolate learned behavior to unseen situations.
We propose a novel method that learns objects' contextualized embeddings with dynamic message passing conditioned on the input natural language.
arXiv Detail & Related papers (2020-09-11T17:35:05Z) - Self-organizing Democratized Learning: Towards Large-scale Distributed
Learning Systems [71.14339738190202]
democratized learning (Dem-AI) lays out a holistic philosophy with underlying principles for building large-scale distributed and democratized machine learning systems.
Inspired by Dem-AI philosophy, a novel distributed learning approach is proposed in this paper.
The proposed algorithms demonstrate better results in the generalization performance of learning models in agents compared to the conventional FL algorithms.
arXiv Detail & Related papers (2020-07-07T08:34:48Z) - Systematic Generalisation through Task Temporal Logic and Deep
Reinforcement Learning [12.136911683449242]
We present a neuro-symbolic framework where a symbolic module transforms TL specifications into a form that helps the training of a DRL agent targeting generalisation.
We study the emergence of systematic learning in different settings and find that the architecture of the convolutional layers is key when generalising to new instructions.
arXiv Detail & Related papers (2020-06-12T09:02:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.