Temporally Extended Goal Recognition in Fully Observable
Non-Deterministic Domain Models
- URL: http://arxiv.org/abs/2306.08680v1
- Date: Wed, 14 Jun 2023 18:02:00 GMT
- Title: Temporally Extended Goal Recognition in Fully Observable
Non-Deterministic Domain Models
- Authors: Ramon Fraga Pereira, Francesco Fuggitti, Felipe Meneguzzi, Giuseppe De
Giacomo
- Abstract summary: Existing approaches assume that goal hypotheses comprise a single conjunctive formula over a single final state.
We focus on temporally extended goals in Fully Observable Non-Deterministic (FOND) planning domain models.
Empirical results show that our approach is accurate in recognizing temporally extended goals in different recognition settings.
- Score: 43.460098744623416
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Goal Recognition is the task of discerning the correct intended goal that an
agent aims to achieve, given a set of goal hypotheses, a domain model, and a
sequence of observations (i.e., a sample of the plan executed in the
environment). Existing approaches assume that goal hypotheses comprise a single
conjunctive formula over a single final state and that the environment dynamics
are deterministic, preventing the recognition of temporally extended goals in
more complex settings. In this paper, we expand goal recognition to temporally
extended goals in Fully Observable Non-Deterministic (FOND) planning domain
models, focusing on goals on finite traces expressed in Linear Temporal Logic
(LTLf) and Pure Past Linear Temporal Logic (PLTLf). We develop the first
approach capable of recognizing goals in such settings and evaluate it using
different LTLf and PLTLf goals over six FOND planning domain models. Empirical
results show that our approach is accurate in recognizing temporally extended
goals in different recognition settings.
Related papers
- Towards Measuring Goal-Directedness in AI Systems [0.0]
A key prerequisite for AI systems pursuing unintended goals is whether they will behave in a coherent and goal-directed manner.
We propose a new family of definitions of the goal-directedness of a policy that analyze whether it is well-modeled as near-optimal for many reward functions.
Our contribution is a definition of goal-directedness that is simpler and more easily computable in order to approach the question of whether AI systems could pursue dangerous goals.
arXiv Detail & Related papers (2024-10-07T01:34:42Z) - Imitating Graph-Based Planning with Goal-Conditioned Policies [72.61631088613048]
We present a self-imitation scheme which distills a subgoal-conditioned policy into the target-goal-conditioned policy.
We empirically show that our method can significantly boost the sample-efficiency of the existing goal-conditioned RL methods.
arXiv Detail & Related papers (2023-03-20T14:51:10Z) - Discrete Factorial Representations as an Abstraction for Goal
Conditioned Reinforcement Learning [99.38163119531745]
We show that applying a discretizing bottleneck can improve performance in goal-conditioned RL setups.
We experimentally prove the expected return on out-of-distribution goals, while still allowing for specifying goals with expressive structure.
arXiv Detail & Related papers (2022-11-01T03:31:43Z) - Goal-Conditioned Q-Learning as Knowledge Distillation [136.79415677706612]
We explore a connection between off-policy reinforcement learning in goal-conditioned settings and knowledge distillation.
We empirically show that this can improve the performance of goal-conditioned off-policy reinforcement learning when the space of goals is high-dimensional.
We also show that this technique can be adapted to allow for efficient learning in the case of multiple simultaneous sparse goals.
arXiv Detail & Related papers (2022-08-28T22:01:10Z) - Adversarial Intrinsic Motivation for Reinforcement Learning [60.322878138199364]
We investigate whether the Wasserstein-1 distance between a policy's state visitation distribution and a target distribution can be utilized effectively for reinforcement learning tasks.
Our approach, termed Adversarial Intrinsic Motivation (AIM), estimates this Wasserstein-1 distance through its dual objective and uses it to compute a supplemental reward function.
arXiv Detail & Related papers (2021-05-27T17:51:34Z) - Recognizing LTLf/PLTLf Goals in Fully Observable Non-Deterministic
Domain Models [26.530274055506453]
Goal Recognition is the task of discerning the correct intended goal that an agent aims to achieve.
We develop a novel approach that is capable of recognizing temporally extended goals.
arXiv Detail & Related papers (2021-03-22T09:46:03Z) - Goal-GAN: Multimodal Trajectory Prediction Based on Goal Position
Estimation [1.20855096102517]
We present Goal-GAN, an interpretable and end-to-end trainable model for human trajectory prediction.
Inspired by human navigation, we model the task of trajectory prediction as an intuitive two-stage process.
arXiv Detail & Related papers (2020-10-02T17:17:45Z) - Maximum Entropy Gain Exploration for Long Horizon Multi-goal
Reinforcement Learning [35.44552072132894]
We argue that a learning agent should set its own intrinsic goals that maximize the entropy of the historical achieved goal distribution.
We show that our strategy achieves an order of magnitude better sample efficiency than the prior state of the art on long-horizon multi-goal tasks.
arXiv Detail & Related papers (2020-07-06T15:36:05Z) - Long-Horizon Visual Planning with Goal-Conditioned Hierarchical
Predictors [124.30562402952319]
The ability to predict and plan into the future is fundamental for agents acting in the world.
Current learning approaches for visual prediction and planning fail on long-horizon tasks.
We propose a framework for visual prediction and planning that is able to overcome both of these limitations.
arXiv Detail & Related papers (2020-06-23T17:58:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.