Upside-Down Reinforcement Learning Can Diverge in Stochastic
Environments With Episodic Resets
- URL: http://arxiv.org/abs/2205.06595v1
- Date: Fri, 13 May 2022 12:43:25 GMT
- Title: Upside-Down Reinforcement Learning Can Diverge in Stochastic
Environments With Episodic Resets
- Authors: Miroslav \v{S}trupl, Francesco Faccio, Dylan R. Ashley, J\"urgen
Schmidhuber, Rupesh Kumar Srivastava
- Abstract summary: Upside-Down Reinforcement Learning (UDRL) is an approach for solving problems that does not require value functions.
Goal-Conditional Supervised Learning (GCSL) optimized a lower bound on goal-reaching performance.
This raises expectations that such algorithms may enjoy guaranteed convergence to the optimal policy in arbitrary environments.
- Score: 4.126347193869613
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Upside-Down Reinforcement Learning (UDRL) is an approach for solving RL
problems that does not require value functions and uses only supervised
learning, where the targets for given inputs in a dataset do not change over
time. Ghosh et al. proved that Goal-Conditional Supervised Learning (GCSL) --
which can be viewed as a simplified version of UDRL -- optimizes a lower bound
on goal-reaching performance. This raises expectations that such algorithms may
enjoy guaranteed convergence to the optimal policy in arbitrary environments,
similar to certain well-known traditional RL algorithms. Here we show that for
a specific episodic UDRL algorithm (eUDRL, including GCSL), this is not the
case, and give the causes of this limitation. To do so, we first introduce a
helpful rewrite of eUDRL as a recursive policy update. This formulation helps
to disprove its convergence to the optimal policy for a wide class of
stochastic environments. Finally, we provide a concrete example of a very
simple environment where eUDRL diverges. Since the primary aim of this paper is
to present a negative result, and the best counterexamples are the simplest
ones, we restrict all discussions to finite (discrete) environments, ignoring
issues of function approximation and limited sample size.
Related papers
- REBEL: Reinforcement Learning via Regressing Relative Rewards [59.68420022466047]
We propose REBEL, a minimalist RL algorithm for the era of generative models.
In theory, we prove that fundamental RL algorithms like Natural Policy Gradient can be seen as variants of REBEL.
We find that REBEL provides a unified approach to language modeling and image generation with stronger or similar performance as PPO and DPO.
arXiv Detail & Related papers (2024-04-25T17:20:45Z) - Offline Imitation Learning from Multiple Baselines with Applications to Compiler Optimization [17.729842629392742]
We study a Reinforcement Learning problem in which we are given a set of trajectories collected with K baseline policies.
The goal is to learn a policy which performs as well as the best combination of baselines on the entire state space.
arXiv Detail & Related papers (2024-03-28T14:34:02Z) - The Effective Horizon Explains Deep RL Performance in Stochastic Environments [21.148001945560075]
Reinforcement learning (RL) theory has largely focused on proving mini complexity sample bounds.
We introduce a new RL algorithm, SQIRL, that iteratively learns a nearoptimal policy by exploring randomly to collect rollouts.
We leverage SQIRL to derive instance-dependent sample complexity bounds for RL that are exponential only in an "effective horizon" look-ahead and on the complexity of the class used for approximation.
arXiv Detail & Related papers (2023-12-13T18:58:56Z) - Hundreds Guide Millions: Adaptive Offline Reinforcement Learning with
Expert Guidance [74.31779732754697]
We propose a novel plug-in approach named Guided Offline RL (GORL)
GORL employs a guiding network, along with only a few expert demonstrations, to adaptively determine the relative importance of the policy improvement and policy constraint for every sample.
Experiments on various environments suggest that GORL can be easily installed on most offline RL algorithms with statistically significant performance improvements.
arXiv Detail & Related papers (2023-09-04T08:59:04Z) - Provably Efficient Offline Goal-Conditioned Reinforcement Learning with
General Function Approximation and Single-Policy Concentrability [11.786486763236104]
Goal-conditioned reinforcement learning (GCRL) refers to learning general-purpose skills that aim to reach diverse goals.
offline GCRL only requires purely pre-collected datasets to perform training tasks.
We show that a modified offline GCRL algorithm is both provably efficient with general function approximation and single-policy concentrability.
arXiv Detail & Related papers (2023-02-07T22:04:55Z) - Human-in-the-loop: Provably Efficient Preference-based Reinforcement
Learning with General Function Approximation [107.54516740713969]
We study human-in-the-loop reinforcement learning (RL) with trajectory preferences.
Instead of receiving a numeric reward at each step, the agent only receives preferences over trajectory pairs from a human overseer.
We propose the first optimistic model-based algorithm for PbRL with general function approximation.
arXiv Detail & Related papers (2022-05-23T09:03:24Z) - Robust Predictable Control [149.71263296079388]
We show that our method achieves much tighter compression than prior methods, achieving up to 5x higher reward than a standard information bottleneck.
We also demonstrate that our method learns policies that are more robust and generalize better to new tasks.
arXiv Detail & Related papers (2021-09-07T17:29:34Z) - Local policy search with Bayesian optimization [73.0364959221845]
Reinforcement learning aims to find an optimal policy by interaction with an environment.
Policy gradients for local search are often obtained from random perturbations.
We develop an algorithm utilizing a probabilistic model of the objective function and its gradient.
arXiv Detail & Related papers (2021-06-22T16:07:02Z) - Online Sub-Sampling for Reinforcement Learning with General Function
Approximation [111.01990889581243]
In this paper, we establish an efficient online sub-sampling framework that measures the information gain of data points collected by an RL algorithm.
For a value-based method with complexity-bounded function class, we show that the policy only needs to be updated for $proptooperatornamepolylog(K)$ times.
In contrast to existing approaches that update the policy for at least $Omega(K)$ times, our approach drastically reduces the number of optimization calls in solving for a policy.
arXiv Detail & Related papers (2021-06-14T07:36:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.