The Role of Exploration for Task Transfer in Reinforcement Learning
- URL: http://arxiv.org/abs/2210.06168v1
- Date: Tue, 11 Oct 2022 01:23:21 GMT
- Title: The Role of Exploration for Task Transfer in Reinforcement Learning
- Authors: Jonathan C Balloch and Julia Kim and and Jessica L Inman and Mark O
Riedl
- Abstract summary: We re-examine the exploration--exploitation trade-off in the context of transfer learning.
In this work, we review reinforcement learning exploration methods, define a taxonomy with which to organize them, analyze these methods' differences in the context of task transfer, and suggest avenues for future investigation.
- Score: 8.817381809671804
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The exploration--exploitation trade-off in reinforcement learning (RL) is a
well-known and much-studied problem that balances greedy action selection with
novel experience, and the study of exploration methods is usually only
considered in the context of learning the optimal policy for a single learning
task. However, in the context of online task transfer, where there is a change
to the task during online operation, we hypothesize that exploration strategies
that anticipate the need to adapt to future tasks can have a pronounced impact
on the efficiency of transfer. As such, we re-examine the
exploration--exploitation trade-off in the context of transfer learning. In
this work, we review reinforcement learning exploration methods, define a
taxonomy with which to organize them, analyze these methods' differences in the
context of task transfer, and suggest avenues for future investigation.
Related papers
- Is Exploration All You Need? Effective Exploration Characteristics for Transfer in Reinforcement Learning [13.322146980081778]
We test eleven popular exploration algorithms on a variety of transfer types -- or novelties'' -- to identify the characteristics that positively affect online transfer learning.
Our analysis shows that some characteristics correlate with improved performance and efficiency across a wide range of transfer tasks, while others only improve transfer performance with respect to specific environment changes.
arXiv Detail & Related papers (2024-04-02T18:45:01Z) - Similarity-based Knowledge Transfer for Cross-Domain Reinforcement
Learning [3.3148826359547523]
We develop a semi-supervised alignment loss to match different spaces with a set of encoder-decoders.
In comparison to prior works, our method does not require data to be aligned, paired or collected by expert policies.
arXiv Detail & Related papers (2023-12-05T19:26:01Z) - Introspective Action Advising for Interpretable Transfer Learning [7.673465837624365]
Transfer learning can be applied in deep reinforcement learning to accelerate the training of a policy in a target task.
We propose an alternative approach to transfer learning between tasks based on action advising, in which a teacher trained in a source task actively guides a student's exploration in a target task.
arXiv Detail & Related papers (2023-06-21T14:53:33Z) - Learning Action Translator for Meta Reinforcement Learning on
Sparse-Reward Tasks [56.63855534940827]
This work introduces a novel objective function to learn an action translator among training tasks.
We theoretically verify that the value of the transferred policy with the action translator can be close to the value of the source policy.
We propose to combine the action translator with context-based meta-RL algorithms for better data collection and more efficient exploration during meta-training.
arXiv Detail & Related papers (2022-07-19T04:58:06Z) - A Survey of Exploration Methods in Reinforcement Learning [64.01676570654234]
Reinforcement learning agents depend crucially on exploration to obtain informative data for the learning process.
In this article, we provide a survey of modern exploration methods in (Sequential) reinforcement learning, as well as a taxonomy of exploration methods.
arXiv Detail & Related papers (2021-09-01T02:36:14Z) - Adaptive Policy Transfer in Reinforcement Learning [9.594432031144715]
We introduce a principled mechanism that can "Adapt-to-Learn", that is adapt the source policy to learn to solve a target task.
We show that the presented method learns to seamlessly combine learning from adaptation and exploration and leads to a robust policy transfer algorithm.
arXiv Detail & Related papers (2021-05-10T22:42:03Z) - Coverage as a Principle for Discovering Transferable Behavior in
Reinforcement Learning [16.12658895065585]
We argue that representation alone is not enough for efficient transfer in challenging domains and explore how to transfer knowledge through behavior.
The behavior of pre-trained policies may be used for solving the task at hand (exploitation) or for collecting useful data to solve the problem (exploration)
arXiv Detail & Related papers (2021-02-24T16:51:02Z) - Unsupervised Transfer Learning for Spatiotemporal Predictive Networks [90.67309545798224]
We study how to transfer knowledge from a zoo of unsupervisedly learned models towards another network.
Our motivation is that models are expected to understand complex dynamics from different sources.
Our approach yields significant improvements on three benchmarks fortemporal prediction, and benefits the target even from less relevant ones.
arXiv Detail & Related papers (2020-09-24T15:40:55Z) - Importance Weighted Policy Learning and Adaptation [89.46467771037054]
We study a complementary approach which is conceptually simple, general, modular and built on top of recent improvements in off-policy learning.
The framework is inspired by ideas from the probabilistic inference literature and combines robust off-policy learning with a behavior prior.
Our approach achieves competitive adaptation performance on hold-out tasks compared to meta reinforcement learning baselines and can scale to complex sparse-reward scenarios.
arXiv Detail & Related papers (2020-09-10T14:16:58Z) - MetaCURE: Meta Reinforcement Learning with Empowerment-Driven
Exploration [52.48362697163477]
Experimental evaluation shows that our meta-RL method significantly outperforms state-of-the-art baselines on sparse-reward tasks.
We model an exploration policy learning problem for meta-RL, which is separated from exploitation policy learning.
We develop a new off-policy meta-RL framework, which efficiently learns separate context-aware exploration and exploitation policies.
arXiv Detail & Related papers (2020-06-15T06:56:18Z) - Curriculum Learning for Reinforcement Learning Domains: A Framework and
Survey [53.73359052511171]
Reinforcement learning (RL) is a popular paradigm for addressing sequential decision tasks in which the agent has only limited environmental feedback.
We present a framework for curriculum learning (CL) in RL, and use it to survey and classify existing CL methods in terms of their assumptions, capabilities, and goals.
arXiv Detail & Related papers (2020-03-10T20:41:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.