A Computational Interface to Translate Strategic Intent from
Unstructured Language in a Low-Data Setting
- URL: http://arxiv.org/abs/2208.08374v2
- Date: Fri, 20 Oct 2023 20:29:43 GMT
- Title: A Computational Interface to Translate Strategic Intent from
Unstructured Language in a Low-Data Setting
- Authors: Pradyumna Tambwekar, Lakshita Dodeja, Nathan Vaska, Wei Xu, Matthew
Gombolay
- Abstract summary: We build a computational interface capable of translating unstructured language strategies into actionable intent in the form of goals and constraints.
We collect a dataset of over 1000 examples, mapping language strategies to the corresponding goals and constraints, and show that our model, trained on this dataset, significantly outperforms human interpreters.
- Score: 7.2466963932212245
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many real-world tasks involve a mixed-initiative setup, wherein humans and AI
systems collaboratively perform a task. While significant work has been
conducted towards enabling humans to specify, through language, exactly how an
agent should complete a task (i.e., low-level specification), prior work lacks
on interpreting the high-level strategic intent of the human commanders.
Parsing strategic intent from language will allow autonomous systems to
independently operate according to the user's plan without frequent guidance or
instruction. In this paper, we build a computational interface capable of
translating unstructured language strategies into actionable intent in the form
of goals and constraints. Leveraging a game environment, we collect a dataset
of over 1000 examples, mapping language strategies to the corresponding goals
and constraints, and show that our model, trained on this dataset,
significantly outperforms human interpreters in inferring strategic intent
(i.e., goals and constraints) from language (p < 0.05). Furthermore, we show
that our model (125M parameters) significantly outperforms ChatGPT for this
task (p < 0.05) in a low-data setting.
Related papers
- Flex: End-to-End Text-Instructed Visual Navigation with Foundation Models [59.892436892964376]
We investigate the minimal data requirements and architectural adaptations necessary to achieve robust closed-loop performance with vision-based control policies.
Our findings are synthesized in Flex (Fly-lexically), a framework that uses pre-trained Vision Language Models (VLMs) as frozen patch-wise feature extractors.
We demonstrate the effectiveness of this approach on quadrotor fly-to-target tasks, where agents trained via behavior cloning successfully generalize to real-world scenes.
arXiv Detail & Related papers (2024-10-16T19:59:31Z) - Learning adaptive planning representations with natural language
guidance [90.24449752926866]
This paper describes Ada, a framework for automatically constructing task-specific planning representations.
Ada interactively learns a library of planner-compatible high-level action abstractions and low-level controllers adapted to a particular domain of planning tasks.
arXiv Detail & Related papers (2023-12-13T23:35:31Z) - Language and Task Arithmetic with Parameter-Efficient Layers for Zero-Shot Summarization [126.96113831681338]
In this paper, we propose to improve zero-shot cross-lingual transfer by composing language or task specialized parameters.
Our method composes language and task PEFT modules via element-wise arithmetic operations to leverage unlabeled data and English labeled data.
arXiv Detail & Related papers (2023-11-15T20:04:58Z) - Efficient Spoken Language Recognition via Multilabel Classification [53.662747523872305]
We show that our models obtain competitive results while being orders of magnitude smaller and faster than current state-of-the-art methods.
Our multilabel strategy is more robust to unseen non-target languages compared to multiclass classification.
arXiv Detail & Related papers (2023-06-02T23:04:19Z) - Language-Conditioned Imitation Learning with Base Skill Priors under Unstructured Data [26.004807291215258]
Language-conditioned robot manipulation aims to develop robots capable of understanding and executing complex tasks.
We propose a general-purpose, language-conditioned approach that combines base skill priors and imitation learning under unstructured data.
We assess our model's performance in both simulated and real-world environments using a zero-shot setting.
arXiv Detail & Related papers (2023-05-30T14:40:38Z) - Grounding Language with Visual Affordances over Unstructured Data [26.92329260907805]
We propose a novel approach to efficiently learn language-conditioned robot skills from unstructured, offline and reset-free data.
We exploit a self-supervised visuo-lingual affordance model, which requires as little as 1% of the total data with language.
We find that our method is capable of completing long-horizon, multi-tier tasks in the real world, while requiring an order of magnitude less data than previous approaches.
arXiv Detail & Related papers (2022-10-04T21:16:48Z) - OCHADAI at SemEval-2022 Task 2: Adversarial Training for Multilingual
Idiomaticity Detection [4.111899441919165]
We propose a multilingual adversarial training model for determining whether a sentence contains an idiomatic expression.
Our model relies on pre-trained contextual representations from different multi-lingual state-of-the-art transformer-based language models.
arXiv Detail & Related papers (2022-06-07T05:52:43Z) - GoalNet: Inferring Conjunctive Goal Predicates from Human Plan
Demonstrations for Robot Instruction Following [15.405156791794191]
Our goal is to enable a robot to learn how to sequence its actions to perform tasks specified as natural language instructions.
We introduce a novel neuro-symbolic model, GoalNet, for contextual and task dependent inference of goal predicates.
GoalNet demonstrates a significant improvement (51%) in the task completion rate in comparison to a state-of-the-art rule-based approach.
arXiv Detail & Related papers (2022-05-14T15:14:40Z) - Context-Aware Language Modeling for Goal-Oriented Dialogue Systems [84.65707332816353]
We formulate goal-oriented dialogue as a partially observed Markov decision process.
We derive a simple and effective method to finetune language models in a goal-aware way.
We evaluate our method on a practical flight-booking task using AirDialogue.
arXiv Detail & Related papers (2022-04-18T17:23:11Z) - Learning Language-Conditioned Robot Behavior from Offline Data and
Crowd-Sourced Annotation [80.29069988090912]
We study the problem of learning a range of vision-based manipulation tasks from a large offline dataset of robot interaction.
We propose to leverage offline robot datasets with crowd-sourced natural language labels.
We find that our approach outperforms both goal-image specifications and language conditioned imitation techniques by more than 25%.
arXiv Detail & Related papers (2021-09-02T17:42:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.