Grounding Language to Entities and Dynamics for Generalization in
Reinforcement Learning
- URL: http://arxiv.org/abs/2101.07393v1
- Date: Tue, 19 Jan 2021 00:59:16 GMT
- Title: Grounding Language to Entities and Dynamics for Generalization in
Reinforcement Learning
- Authors: H. J. Austin Wang and Karthik Narasimhan
- Abstract summary: We consider the problem of leveraging textual descriptions to improve generalization of control policies to new scenarios.
We develop a new model, EMMA, which uses a multi-modal entity-conditioned attention module.
EMMA is end-to-end differentiable and can learn a latent grounding of entities and dynamics from text to observations.
- Score: 20.43004852346133
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we consider the problem of leveraging textual descriptions to
improve generalization of control policies to new scenarios. Unlike prior work
in this space, we do not assume access to any form of prior knowledge
connecting text and state observations, and learn both symbol grounding and
control policy simultaneously. This is challenging due to a lack of concrete
supervision, and incorrect groundings can result in worse performance than
policies that do not use the text at all. We develop a new model, EMMA (Entity
Mapper with Multi-modal Attention) which uses a multi-modal entity-conditioned
attention module that allows for selective focus over relevant sentences in the
manual for each entity in the environment. EMMA is end-to-end differentiable
and can learn a latent grounding of entities and dynamics from text to
observations using environment rewards as the only source of supervision. To
empirically test our model, we design a new framework of 1320 games and collect
text manuals with free-form natural language via crowd-sourcing. We demonstrate
that EMMA achieves successful zero-shot generalization to unseen games with new
dynamics, obtaining significantly higher rewards compared to multiple
baselines. The grounding acquired by EMMA is also robust to noisy descriptions
and linguistic variation.
Related papers
- Text-to-Decision Agent: Learning Generalist Policies from Natural Language Supervision [36.643102023506614]
We propose Text-to-Decision Agent (T2DA), a framework that supervises generalist policy learning with natural language.
We show that T2DA facilitates high-capacity zero-shot generalization and outperforms various types of baselines.
arXiv Detail & Related papers (2025-04-21T12:00:20Z) - LangSuitE: Planning, Controlling and Interacting with Large Language Models in Embodied Text Environments [70.91258869156353]
We introduce LangSuitE, a versatile and simulation-free testbed featuring 6 representative embodied tasks in textual embodied worlds.
Compared with previous LLM-based testbeds, LangSuitE offers adaptability to diverse environments without multiple simulation engines.
We devise a novel chain-of-thought (CoT) schema, EmMem, which summarizes embodied states w.r.t. history information.
arXiv Detail & Related papers (2024-06-24T03:36:29Z) - Scalable Language Model with Generalized Continual Learning [58.700439919096155]
The Joint Adaptive Re-ization (JARe) is integrated with Dynamic Task-related Knowledge Retrieval (DTKR) to enable adaptive adjustment of language models based on specific downstream tasks.
Our method demonstrates state-of-the-art performance on diverse backbones and benchmarks, achieving effective continual learning in both full-set and few-shot scenarios with minimal forgetting.
arXiv Detail & Related papers (2024-04-11T04:22:15Z) - Text2Data: Low-Resource Data Generation with Textual Control [104.38011760992637]
Natural language serves as a common and straightforward control signal for humans to interact seamlessly with machines.
We propose Text2Data, a novel approach that utilizes unlabeled data to understand the underlying data distribution through an unsupervised diffusion model.
It undergoes controllable finetuning via a novel constraint optimization-based learning objective that ensures controllability and effectively counteracts catastrophic forgetting.
arXiv Detail & Related papers (2024-02-08T03:41:39Z) - LanGWM: Language Grounded World Model [24.86620763902546]
We focus on learning language-grounded visual features to enhance the world model learning.
Our proposed technique of explicit language-grounded visual representation learning has the potential to improve models for human-robot interaction.
arXiv Detail & Related papers (2023-11-29T12:41:55Z) - Adapting Large Language Models to Domains via Reading Comprehension [86.24451681746676]
We explore how continued pre-training on domain-specific corpora influences large language models.
We show that training on the raw corpora endows the model with domain knowledge, but drastically hurts its ability for question answering.
We propose a simple method for transforming raw corpora into reading comprehension texts.
arXiv Detail & Related papers (2023-09-18T07:17:52Z) - Learning Symbolic Rules over Abstract Meaning Representations for
Textual Reinforcement Learning [63.148199057487226]
We propose a modular, NEuroSymbolic Textual Agent (NESTA) that combines a generic semantic generalization with a rule induction system to learn interpretable rules as policies.
Our experiments show that the proposed NESTA method outperforms deep reinforcement learning-based techniques by achieving better to unseen test games and learning from fewer training interactions.
arXiv Detail & Related papers (2023-07-05T23:21:05Z) - Dependency Induction Through the Lens of Visual Perception [81.91502968815746]
We propose an unsupervised grammar induction model that leverages word concreteness and a structural vision-based to jointly learn constituency-structure and dependency-structure grammars.
Our experiments show that the proposed extension outperforms the current state-of-the-art visually grounded models in constituency parsing even with a smaller grammar size.
arXiv Detail & Related papers (2021-09-20T18:40:37Z) - KELM: Knowledge Enhanced Pre-Trained Language Representations with
Message Passing on Hierarchical Relational Graphs [26.557447199727758]
We propose a novel knowledge-aware language model framework based on fine-tuning process.
Our model can efficiently incorporate world knowledge from KGs into existing language models such as BERT.
arXiv Detail & Related papers (2021-09-09T12:39:17Z) - Zero-Shot Compositional Policy Learning via Language Grounding [13.45138913186308]
Humans can adapt to new tasks quickly by leveraging prior knowledge about the world such as language descriptions.
We introduce a new research platform BabyAI++ in which the dynamics of environments are disentangled from visual appearance.
We find that current language-guided RL/IL techniques overfit to the training environments and suffer from a huge performance drop when facing unseen combinations.
arXiv Detail & Related papers (2020-04-15T16:58:19Z) - Contextual Policy Transfer in Reinforcement Learning Domains via Deep
Mixtures-of-Experts [24.489002406693128]
We introduce a novel mixture-of-experts formulation for learning state-dependent beliefs over source task dynamics.
We show how this model can be incorporated into standard policy reuse frameworks.
arXiv Detail & Related papers (2020-02-29T07:58:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.