Recognizing LTLf/PLTLf Goals in Fully Observable Non-Deterministic
Domain Models
- URL: http://arxiv.org/abs/2103.11692v1
- Date: Mon, 22 Mar 2021 09:46:03 GMT
- Title: Recognizing LTLf/PLTLf Goals in Fully Observable Non-Deterministic
Domain Models
- Authors: Ramon Fraga Pereira, Francesco Fuggitti, and Giuseppe De Giacomo
- Abstract summary: Goal Recognition is the task of discerning the correct intended goal that an agent aims to achieve.
We develop a novel approach that is capable of recognizing temporally extended goals.
- Score: 26.530274055506453
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Goal Recognition is the task of discerning the correct intended goal that an
agent aims to achieve, given a set of possible goals, a domain model, and a
sequence of observations as a sample of the plan being executed in the
environment. Existing approaches assume that the possible goals are formalized
as a conjunction in deterministic settings. In this paper, we develop a novel
approach that is capable of recognizing temporally extended goals in Fully
Observable Non-Deterministic (FOND) planning domain models, focusing on goals
on finite traces expressed in Linear Temporal Logic (LTLf) and (Pure) Past
Linear Temporal Logic (PLTLf). We empirically evaluate our goal recognition
approach using different LTLf and PLTLf goals over six common FOND planning
domain models, and show that our approach is accurate to recognize temporally
extended goals at several levels of observability.
Related papers
- Directed Exploration in Reinforcement Learning from Linear Temporal Logic [59.707408697394534]
Linear temporal logic (LTL) is a powerful language for task specification in reinforcement learning.
We show that the synthesized reward signal remains fundamentally sparse, making exploration challenging.
We show how better exploration can be achieved by further leveraging the specification and casting its corresponding Limit Deterministic B"uchi Automaton (LDBA) as a Markov reward process.
arXiv Detail & Related papers (2024-08-18T14:25:44Z) - Temporally Extended Goal Recognition in Fully Observable
Non-Deterministic Domain Models [43.460098744623416]
Existing approaches assume that goal hypotheses comprise a single conjunctive formula over a single final state.
We focus on temporally extended goals in Fully Observable Non-Deterministic (FOND) planning domain models.
Empirical results show that our approach is accurate in recognizing temporally extended goals in different recognition settings.
arXiv Detail & Related papers (2023-06-14T18:02:00Z) - Imitating Graph-Based Planning with Goal-Conditioned Policies [72.61631088613048]
We present a self-imitation scheme which distills a subgoal-conditioned policy into the target-goal-conditioned policy.
We empirically show that our method can significantly boost the sample-efficiency of the existing goal-conditioned RL methods.
arXiv Detail & Related papers (2023-03-20T14:51:10Z) - Discrete Factorial Representations as an Abstraction for Goal
Conditioned Reinforcement Learning [99.38163119531745]
We show that applying a discretizing bottleneck can improve performance in goal-conditioned RL setups.
We experimentally prove the expected return on out-of-distribution goals, while still allowing for specifying goals with expressive structure.
arXiv Detail & Related papers (2022-11-01T03:31:43Z) - C-Planning: An Automatic Curriculum for Learning Goal-Reaching Tasks [133.40619754674066]
Goal-conditioned reinforcement learning can solve tasks in a wide range of domains, including navigation and manipulation.
We propose the distant goal-reaching task by using search at training time to automatically generate intermediate states.
E-step corresponds to planning an optimal sequence of waypoints using graph search, while the M-step aims to learn a goal-conditioned policy to reach those waypoints.
arXiv Detail & Related papers (2021-10-22T22:05:31Z) - Goal-GAN: Multimodal Trajectory Prediction Based on Goal Position
Estimation [1.20855096102517]
We present Goal-GAN, an interpretable and end-to-end trainable model for human trajectory prediction.
Inspired by human navigation, we model the task of trajectory prediction as an intuitive two-stage process.
arXiv Detail & Related papers (2020-10-02T17:17:45Z) - Maximum Entropy Gain Exploration for Long Horizon Multi-goal
Reinforcement Learning [35.44552072132894]
We argue that a learning agent should set its own intrinsic goals that maximize the entropy of the historical achieved goal distribution.
We show that our strategy achieves an order of magnitude better sample efficiency than the prior state of the art on long-horizon multi-goal tasks.
arXiv Detail & Related papers (2020-07-06T15:36:05Z) - Long-Horizon Visual Planning with Goal-Conditioned Hierarchical
Predictors [124.30562402952319]
The ability to predict and plan into the future is fundamental for agents acting in the world.
Current learning approaches for visual prediction and planning fail on long-horizon tasks.
We propose a framework for visual prediction and planning that is able to overcome both of these limitations.
arXiv Detail & Related papers (2020-06-23T17:58:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.