Reinforcement Learning for Omega-Regular Specifications on
Continuous-Time MDP
- URL: http://arxiv.org/abs/2303.09528v1
- Date: Thu, 16 Mar 2023 17:45:38 GMT
- Title: Reinforcement Learning for Omega-Regular Specifications on
Continuous-Time MDP
- Authors: Amin Falah, Shibashis Guha, Ashutosh Trivedi
- Abstract summary: Continuous-time Markov decision processes (CTMDPs) are canonical models to express sequential decision-making under dense-time environments.
We present an approach enabling correct translation to scalar reward signals for CTMDPs.
- Score: 1.8262547855491456
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Continuous-time Markov decision processes (CTMDPs) are canonical models to
express sequential decision-making under dense-time and stochastic
environments. When the stochastic evolution of the environment is only
available via sampling, model-free reinforcement learning (RL) is the
algorithm-of-choice to compute optimal decision sequence. RL, on the other
hand, requires the learning objective to be encoded as scalar reward signals.
Since doing such translations manually is both tedious and error-prone, a
number of techniques have been proposed to translate high-level objectives
(expressed in logic or automata formalism) to scalar rewards for discrete-time
Markov decision processes (MDPs). Unfortunately, no automatic translation
exists for CTMDPs.
We consider CTMDP environments against the learning objectives expressed as
omega-regular languages. Omega-regular languages generalize regular languages
to infinite-horizon specifications and can express properties given in popular
linear-time logic LTL. To accommodate the dense-time nature of CTMDPs, we
consider two different semantics of omega-regular objectives: 1) satisfaction
semantics where the goal of the learner is to maximize the probability of
spending positive time in the good states, and 2) expectation semantics where
the goal of the learner is to optimize the long-run expected average time spent
in the ``good states" of the automaton. We present an approach enabling correct
translation to scalar reward signals that can be readily used by off-the-shelf
RL algorithms for CTMDPs. We demonstrate the effectiveness of the proposed
algorithms by evaluating it on some popular CTMDP benchmarks with omega-regular
objectives.
Related papers
- Directed Exploration in Reinforcement Learning from Linear Temporal Logic [59.707408697394534]
Linear temporal logic (LTL) is a powerful language for task specification in reinforcement learning.
We show that the synthesized reward signal remains fundamentally sparse, making exploration challenging.
We show how better exploration can be achieved by further leveraging the specification and casting its corresponding Limit Deterministic B"uchi Automaton (LDBA) as a Markov reward process.
arXiv Detail & Related papers (2024-08-18T14:25:44Z) - Entropy-Regularized Token-Level Policy Optimization for Language Agent Reinforcement [67.1393112206885]
Large Language Models (LLMs) have shown promise as intelligent agents in interactive decision-making tasks.
We introduce Entropy-Regularized Token-level Policy Optimization (ETPO), an entropy-augmented RL method tailored for optimizing LLMs at the token level.
We assess the effectiveness of ETPO within a simulated environment that models data science code generation as a series of multi-step interactive tasks.
arXiv Detail & Related papers (2024-02-09T07:45:26Z) - A PAC Learning Algorithm for LTL and Omega-regular Objectives in MDPs [5.946838062187346]
We introduce a model-based probably approximately correct (PAC) learning algorithm for omega-regular objectives in decision processes (MDPs)
We prove that our algorithm only requires a number of samples to perform experiments which confirm our theory.
arXiv Detail & Related papers (2023-10-18T18:33:41Z) - Large Language Models as General Pattern Machines [64.75501424160748]
We show that pre-trained large language models (LLMs) are capable of autoregressively completing complex token sequences.
Surprisingly, pattern completion proficiency can be partially retained even when the sequences are expressed using tokens randomly sampled from the vocabulary.
In this work, we investigate how these zero-shot capabilities may be applied to problems in robotics.
arXiv Detail & Related papers (2023-07-10T17:32:13Z) - Quick Adaptive Ternary Segmentation: An Efficient Decoding Procedure For
Hidden Markov Models [70.26374282390401]
Decoding the original signal (i.e., hidden chain) from the noisy observations is one of the main goals in nearly all HMM based data analyses.
We present Quick Adaptive Ternary (QATS), a divide-and-conquer procedure which decodes the hidden sequence in polylogarithmic computational complexity.
arXiv Detail & Related papers (2023-05-29T19:37:48Z) - Stopping Criteria for Value Iteration on Stochastic Games with
Quantitative Objectives [0.0]
A classic solution technique for Markov decision processes (MDP) and games (SG) is value (VI)
In this paper, we provide the first stopping criteria for VI on SG with total reward and mean payoff, yielding the first anytime algorithms in these settings.
arXiv Detail & Related papers (2023-04-19T19:09:55Z) - Universal Learning Waveform Selection Strategies for Adaptive Target
Tracking [42.4297040396286]
This work develops a sequential waveform selection scheme which achieves Bellman optimality in any radar scene.
An algorithm based on a multi-alphabet version of the Context-Tree Weighting (CTW) method can be used to optimally solve a broad class of waveform-agile tracking problems.
arXiv Detail & Related papers (2022-02-10T19:21:03Z) - Modular Deep Reinforcement Learning for Continuous Motion Planning with
Temporal Logic [59.94347858883343]
This paper investigates the motion planning of autonomous dynamical systems modeled by Markov decision processes (MDP)
The novelty is to design an embedded product MDP (EP-MDP) between the LDGBA and the MDP.
The proposed LDGBA-based reward shaping and discounting schemes for the model-free reinforcement learning (RL) only depend on the EP-MDP states.
arXiv Detail & Related papers (2021-02-24T01:11:25Z) - Certified Reinforcement Learning with Logic Guidance [78.2286146954051]
We propose a model-free RL algorithm that enables the use of Linear Temporal Logic (LTL) to formulate a goal for unknown continuous-state/action Markov Decision Processes (MDPs)
The algorithm is guaranteed to synthesise a control policy whose traces satisfy the specification with maximal probability.
arXiv Detail & Related papers (2019-02-02T20:09:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.