Zero-Shot Instruction Following in RL via Structured LTL Representations
- URL: http://arxiv.org/abs/2602.14344v1
- Date: Sun, 15 Feb 2026 23:22:50 GMT
- Title: Zero-Shot Instruction Following in RL via Structured LTL Representations
- Authors: Mathias Jackermeier, Mattia Giuri, Jacques Cloete, Alessandro Abate,
- Abstract summary: We study instruction following in multi-task reinforcement learning, where an agent must zero-shot execute novel tasks not seen during training.<n>In this setting, linear temporal logic has recently been adopted as a powerful framework for specifying structured, temporally extended tasks.<n>While existing approaches successfully train generalist policies, they often struggle to effectively capture the rich logical and temporal structure inherent in specifications.
- Score: 50.41415009303967
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study instruction following in multi-task reinforcement learning, where an agent must zero-shot execute novel tasks not seen during training. In this setting, linear temporal logic (LTL) has recently been adopted as a powerful framework for specifying structured, temporally extended tasks. While existing approaches successfully train generalist policies, they often struggle to effectively capture the rich logical and temporal structure inherent in LTL specifications. In this work, we address these concerns with a novel approach to learn structured task representations that facilitate training and generalisation. Our method conditions the policy on sequences of Boolean formulae constructed from a finite automaton of the task. We propose a hierarchical neural architecture to encode the logical structure of these formulae, and introduce an attention mechanism that enables the policy to reason about future subgoals. Experiments in a variety of complex environments demonstrate the strong generalisation capabilities and superior performance of our approach.
Related papers
- Semantically Labelled Automata for Multi-Task Reinforcement Learning with LTL Instructions [61.479946958462754]
We study multi-task reinforcement learning (RL), a setting in which an agent learns a single, universal policy.<n>We present a novel task embedding technique leveraging a new generation of semantic translations-to-automata.
arXiv Detail & Related papers (2026-02-06T14:46:27Z) - PlatoLTL: Learning to Generalize Across Symbols in LTL Instructions for Multi-Task RL [55.58188508467081]
linear temporal logic (LTL) is a powerful formalism for specifying structured, temporally extended tasks to RL agents.<n>We present PlatoLTL, a novel approach that enables policies to zero-shot generalize not only compositionally across formula structures, but also parametrically across propositions.
arXiv Detail & Related papers (2026-01-30T12:11:55Z) - Zero-Shot Instruction Following in RL via Structured LTL Representations [54.08661695738909]
Linear temporal logic (LTL) is a compelling framework for specifying complex, structured tasks for reinforcement learning (RL) agents.<n>Recent work has shown that interpreting instructions as finite automata, which can be seen as high-level programs monitoring task progress, enables learning a single generalist policy capable of executing arbitrary instructions at test time.<n>We propose a novel approach to learning a multi-task policy for following arbitrary instructions that addresses this shortcoming.
arXiv Detail & Related papers (2025-12-02T10:44:51Z) - DeepLTL: Learning to Efficiently Satisfy Complex LTL Specifications for Multi-Task RL [59.01527054553122]
Linear temporal logic (LTL) has recently been adopted as a powerful formalism for specifying complex, temporally extended tasks.<n>Existing approaches suffer from several shortcomings.<n>We propose a novel learning approach to address these concerns.
arXiv Detail & Related papers (2024-10-06T21:30:38Z) - Neurosymbolic AI for Enhancing Instructability in Generative AI [7.4348066967005275]
Generative AI has transformed content creation across text, images, and music, showcasing capabilities in following instructions through prompting.
This article explores why neurosymbolic AI offers a better path to enhance the instructability of Large Language Models (LLMs)
We show that neurosymbolic approach enhances the reliability and context-awareness of task execution, enabling LLMs to dynamically interpret and respond to a wider range of instructional contexts with greater precision and flexibility.
arXiv Detail & Related papers (2024-07-26T13:15:50Z) - In a Nutshell, the Human Asked for This: Latent Goals for Following
Temporal Specifications [16.9640514047609]
We address the problem of building agents whose goal is to satisfy out-of distribution (OOD) multi-task instructions expressed in temporal logic (TL)
Recent works provided evidence that the deep learning architecture is a key feature when teaching a DRL agent to solve OOD tasks in TL.
We present a novel deep learning architecture that induces agents to generate latent representations of their current goal given both the human instruction and the current observation from the environment.
arXiv Detail & Related papers (2021-10-18T16:53:31Z) - Multi-Agent Reinforcement Learning with Temporal Logic Specifications [65.79056365594654]
We study the problem of learning to satisfy temporal logic specifications with a group of agents in an unknown environment.
We develop the first multi-agent reinforcement learning technique for temporal logic specifications.
We provide correctness and convergence guarantees for our main algorithm.
arXiv Detail & Related papers (2021-02-01T01:13:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.