Explaining Multi-stage Tasks by Learning Temporal Logic Formulas from
Suboptimal Demonstrations
- URL: http://arxiv.org/abs/2006.02411v1
- Date: Wed, 3 Jun 2020 17:40:14 GMT
- Title: Explaining Multi-stage Tasks by Learning Temporal Logic Formulas from
Suboptimal Demonstrations
- Authors: Glen Chou, Necmiye Ozay, Dmitry Berenson
- Abstract summary: We present a method for learning multi-stage tasks from demonstrations by learning the logical structure and atomic propositions of a consistent linear temporal logic (LTL) formula.
The learner is given successful but potentially suboptimal demonstrations, where the demonstrator is optimizing a cost function while satisfying the formula, and the cost function is uncertain to the learner.
Our algorithm uses the Karush-Kuhn-Tucker (KKT) optimality conditions of the demonstrations together with a counter-example-guided falsification strategy to learn the atomic proposition parameters.
- Score: 6.950510860295866
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a method for learning multi-stage tasks from demonstrations by
learning the logical structure and atomic propositions of a consistent linear
temporal logic (LTL) formula. The learner is given successful but potentially
suboptimal demonstrations, where the demonstrator is optimizing a cost function
while satisfying the LTL formula, and the cost function is uncertain to the
learner. Our algorithm uses the Karush-Kuhn-Tucker (KKT) optimality conditions
of the demonstrations together with a counterexample-guided falsification
strategy to learn the atomic proposition parameters and logical structure of
the LTL formula, respectively. We provide theoretical guarantees on the
conservativeness of the recovered atomic proposition sets, as well as
completeness in the search for finding an LTL formula consistent with the
demonstrations. We evaluate our method on high-dimensional nonlinear systems by
learning LTL formulas explaining multi-stage tasks on 7-DOF arm and quadrotor
systems and show that it outperforms competing methods for learning LTL
formulas from positive examples.
Related papers
- Directed Exploration in Reinforcement Learning from Linear Temporal Logic [59.707408697394534]
Linear temporal logic (LTL) is a powerful language for task specification in reinforcement learning.
We show that the synthesized reward signal remains fundamentally sparse, making exploration challenging.
We show how better exploration can be achieved by further leveraging the specification and casting its corresponding Limit Deterministic B"uchi Automaton (LDBA) as a Markov reward process.
arXiv Detail & Related papers (2024-08-18T14:25:44Z) - On-the-fly Synthesis for LTL over Finite Traces: An Efficient Approach that Counts [20.14001970300658]
We present an on-the-fly framework for Linear Temporal Logic over finite traces (LTLf) based on top-down deterministic automata construction.
arXiv Detail & Related papers (2024-08-14T06:52:58Z) - TLINet: Differentiable Neural Network Temporal Logic Inference [10.36033062385604]
This paper introduces TLINet, a neural-symbolic framework for learning STL formulas.
In contrast to existing approaches, we introduce approximation methods for max operator designed specifically for temporal logic-based gradient techniques.
Our framework not only learns the structure but also the parameters of STL formulas, allowing flexible combinations of operators and various logical structures.
arXiv Detail & Related papers (2024-05-03T16:38:14Z) - Integrating Explanations in Learning LTL Specifications from Demonstrations [6.070833893646998]
This paper investigates whether recent advances in Large Language Models (LLMs) can assist in translating human explanations into a format that can robustly support learning Linear Temporal Logic (LTL) from demonstrations.
We present a principled approach combining LLMs and optimization-based methods to faithfully translate human explanations and demonstrations into specifications.
arXiv Detail & Related papers (2024-04-03T17:09:00Z) - The Common Stability Mechanism behind most Self-Supervised Learning
Approaches [64.40701218561921]
We provide a framework to explain the stability mechanism of different self-supervised learning techniques.
We discuss the working mechanism of contrastive techniques like SimCLR, non-contrastive techniques like BYOL, SWAV, SimSiam, Barlow Twins, and DINO.
We formulate different hypotheses and test them using the Imagenet100 dataset.
arXiv Detail & Related papers (2024-02-22T20:36:24Z) - Faithful Explanations of Black-box NLP Models Using LLM-generated
Counterfactuals [67.64770842323966]
Causal explanations of predictions of NLP systems are essential to ensure safety and establish trust.
Existing methods often fall short of explaining model predictions effectively or efficiently.
We propose two approaches for counterfactual (CF) approximation.
arXiv Detail & Related papers (2023-10-01T07:31:04Z) - Linear Temporal Logic Modulo Theories over Finite Traces (Extended
Version) [72.38188258853155]
This paper studies Linear Temporal Logic over Finite Traces (LTLf)
proposition letters are replaced with first-order formulas interpreted over arbitrary theories.
The resulting logic, called Satisfiability Modulo Theories (LTLfMT), is semi-decidable.
arXiv Detail & Related papers (2022-04-28T17:57:33Z) - Reinforcement Learning from Partial Observation: Linear Function Approximation with Provable Sample Efficiency [111.83670279016599]
We study reinforcement learning for partially observed decision processes (POMDPs) with infinite observation and state spaces.
We make the first attempt at partial observability and function approximation for a class of POMDPs with a linear structure.
arXiv Detail & Related papers (2022-04-20T21:15:38Z) - Learning Model Checking and the Kernel Trick for Signal Temporal Logic
on Stochastic Processes [1.2708506121941319]
We introduce a similarity function on formulae of signal temporal logic (STL)
The corresponding kernel trick allows us to circumvent the complicated process of feature extraction.
We demonstrate this consequence and its advantages on the task of predicting (quantitative) satisfaction of STL formulae.
arXiv Detail & Related papers (2022-01-24T19:36:11Z) - Formal Synthesis of Lyapunov Neural Networks [61.79595926825511]
We propose an automatic and formally sound method for synthesising Lyapunov functions.
We employ a counterexample-guided approach where a numerical learner and a symbolic verifier interact to construct provably correct Lyapunov neural networks.
Our method synthesises Lyapunov functions faster and over wider spatial domains than the alternatives, yet providing stronger or equal guarantees.
arXiv Detail & Related papers (2020-03-19T17:21:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.