Draw Me a Flower: Grounding Formal Abstract Structures Stated in
Informal Natural Language
- URL: http://arxiv.org/abs/2106.14321v1
- Date: Sun, 27 Jun 2021 21:11:16 GMT
- Title: Draw Me a Flower: Grounding Formal Abstract Structures Stated in
Informal Natural Language
- Authors: Royi Lachmy, Valentina Pyatkin, Reut Tsarfaty
- Abstract summary: We develop the Hexagons referential game, where players describe increasingly complex images on a two-dimensional Hexagons board.
Using this game we collected the Hexagons dataset, which consists of 164 images and over 3000 naturally-occurring instructions.
Results of our baseline models on an instruction-to-execution task derived from the Hexagons dataset confirm that higher-level abstractions in NL are indeed more challenging for current systems to process.
- Score: 6.900102922776184
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Forming and interpreting abstraction is a core process in human
communication. In particular, when giving and performing complex instructions
stated in natural language (NL), people may naturally evoke abstract constructs
such as objects, loops, conditions and functions to convey their intentions in
an efficient and precise way. Yet, interpreting and grounding abstraction
stated in NL has not been systematically studied in NLP/AI. To elicit
naturally-occurring abstractions in NL we develop the Hexagons referential
game, where players describe increasingly complex images on a two-dimensional
Hexagons board, and other players need to follow these instructions to recreate
the images. Using this game we collected the Hexagons dataset, which consists
of 164 images and over 3000 naturally-occurring instructions, rich with diverse
abstractions. Results of our baseline models on an instruction-to-execution
task derived from the Hexagons dataset confirm that higher-level abstractions
in NL are indeed more challenging for current systems to process. Thus, this
dataset exposes a new and challenging dimension for grounded semantic parsing,
and we propose it for the community as a future benchmark to explore more
sophisticated and high-level communication within NLP applications.
Related papers
- VisualPredicator: Learning Abstract World Models with Neuro-Symbolic Predicates for Robot Planning [86.59849798539312]
We present Neuro-Symbolic Predicates, a first-order abstraction language that combines the strengths of symbolic and neural knowledge representations.
We show that our approach offers better sample complexity, stronger out-of-distribution generalization, and improved interpretability.
arXiv Detail & Related papers (2024-10-30T16:11:05Z) - How to Handle Sketch-Abstraction in Sketch-Based Image Retrieval? [120.49126407479717]
We propose a sketch-based image retrieval framework capable of handling sketch abstraction at varied levels.
For granularity-level abstraction understanding, we dictate that the retrieval model should not treat all abstraction-levels equally.
Our Acc.@q loss uniquely allows a sketch to narrow/broaden its focus in terms of how stringent the evaluation should be.
arXiv Detail & Related papers (2024-03-11T23:08:29Z) - Learning with Language-Guided State Abstractions [58.199148890064826]
Generalizable policy learning in high-dimensional observation spaces is facilitated by well-designed state representations.
Our method, LGA, uses a combination of natural language supervision and background knowledge from language models to automatically build state representations tailored to unseen tasks.
Experiments on simulated robotic tasks show that LGA yields state abstractions similar to those designed by humans, but in a fraction of the time.
arXiv Detail & Related papers (2024-02-28T23:57:04Z) - Emergence and Function of Abstract Representations in Self-Supervised
Transformers [0.0]
We study the inner workings of small-scale transformers trained to reconstruct partially masked visual scenes.
We show that the network develops intermediate abstract representations, or abstractions, that encode all semantic features of the dataset.
Using precise manipulation experiments, we demonstrate that abstractions are central to the network's decision-making process.
arXiv Detail & Related papers (2023-12-08T20:47:15Z) - AbsPyramid: Benchmarking the Abstraction Ability of Language Models with a Unified Entailment Graph [62.685920585838616]
abstraction ability is essential in human intelligence, which remains under-explored in language models.
We present AbsPyramid, a unified entailment graph of 221K textual descriptions of abstraction knowledge.
arXiv Detail & Related papers (2023-11-15T18:11:23Z) - tagE: Enabling an Embodied Agent to Understand Human Instructions [3.943519623674811]
We introduce a novel system known as task and argument grounding for Embodied agents (tagE)
At its core, our system employs an inventive neural network model designed to extract a series of tasks from complex task instructions expressed in natural language.
Our proposed model adopts an encoder-decoder framework enriched with nested decoding to effectively extract tasks and their corresponding arguments from these intricate instructions.
arXiv Detail & Related papers (2023-10-24T08:17:48Z) - Semantic Exploration from Language Abstractions and Pretrained
Representations [23.02024937564099]
Effective exploration is a challenge in reinforcement learning (RL)
We define novelty using semantically meaningful state abstractions.
We evaluate vision-language representations, pretrained on natural image captioning datasets.
arXiv Detail & Related papers (2022-04-08T17:08:00Z) - Generating Instructions at Different Levels of Abstraction [61.70390291746106]
We show how to generate building instructions at different levels of abstraction in Minecraft.
A crowdsourcing evaluation shows that the choice of abstraction level matters to users.
arXiv Detail & Related papers (2020-10-08T13:56:09Z) - Improving Disentangled Text Representation Learning with
Information-Theoretic Guidance [99.68851329919858]
discrete nature of natural language makes disentangling of textual representations more challenging.
Inspired by information theory, we propose a novel method that effectively manifests disentangled representations of text.
Experiments on both conditional text generation and text-style transfer demonstrate the high quality of our disentangled representation.
arXiv Detail & Related papers (2020-06-01T03:36:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.