Learning from Guided Play: Improving Exploration for Adversarial
Imitation Learning with Simple Auxiliary Tasks
- URL: http://arxiv.org/abs/2301.00051v2
- Date: Thu, 12 Oct 2023 21:47:53 GMT
- Title: Learning from Guided Play: Improving Exploration for Adversarial
Imitation Learning with Simple Auxiliary Tasks
- Authors: Trevor Ablett, Bryan Chan, Jonathan Kelly
- Abstract summary: We show that the standard, naive approach to exploration can manifest as a suboptimal local maximum.
We present Learning from Guided Play (LfGP), a framework in which we leverage expert demonstrations of multiple exploratory, auxiliary tasks.
- Score: 8.320969283401233
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Adversarial imitation learning (AIL) has become a popular alternative to
supervised imitation learning that reduces the distribution shift suffered by
the latter. However, AIL requires effective exploration during an online
reinforcement learning phase. In this work, we show that the standard, naive
approach to exploration can manifest as a suboptimal local maximum if a policy
learned with AIL sufficiently matches the expert distribution without fully
learning the desired task. This can be particularly catastrophic for
manipulation tasks, where the difference between an expert and a non-expert
state-action pair is often subtle. We present Learning from Guided Play (LfGP),
a framework in which we leverage expert demonstrations of multiple exploratory,
auxiliary tasks in addition to a main task. The addition of these auxiliary
tasks forces the agent to explore states and actions that standard AIL may
learn to ignore. Additionally, this particular formulation allows for the
reusability of expert data between main tasks. Our experimental results in a
challenging multitask robotic manipulation domain indicate that LfGP
significantly outperforms both AIL and behaviour cloning, while also being more
expert sample efficient than these baselines. To explain this performance gap,
we provide further analysis of a toy problem that highlights the coupling
between a local maximum and poor exploration, and also visualize the
differences between the learned models from AIL and LfGP.
Related papers
- Sample Efficient Myopic Exploration Through Multitask Reinforcement
Learning with Diverse Tasks [53.44714413181162]
This paper shows that when an agent is trained on a sufficiently diverse set of tasks, a generic policy-sharing algorithm with myopic exploration design can be sample-efficient.
To the best of our knowledge, this is the first theoretical demonstration of the "exploration benefits" of MTRL.
arXiv Detail & Related papers (2024-03-03T22:57:44Z) - RLIF: Interactive Imitation Learning as Reinforcement Learning [56.997263135104504]
We show how off-policy reinforcement learning can enable improved performance under assumptions that are similar but potentially even more practical than those of interactive imitation learning.
Our proposed method uses reinforcement learning with user intervention signals themselves as rewards.
This relaxes the assumption that intervening experts in interactive imitation learning should be near-optimal and enables the algorithm to learn behaviors that improve over the potential suboptimal human expert.
arXiv Detail & Related papers (2023-11-21T21:05:21Z) - A Simple Solution for Offline Imitation from Observations and Examples
with Possibly Incomplete Trajectories [122.11358440078581]
offline imitation is useful in real-world scenarios where arbitrary interactions are costly and expert actions are unavailable.
We propose Trajectory-Aware Learning from Observations (TAILO) to solve MDPs where only task-specific expert states and task-agnostic non-expert state-action pairs are available.
arXiv Detail & Related papers (2023-11-02T15:41:09Z) - Learning from Pixels with Expert Observations [0.0]
This paper presents a new approach that uses expert observations for learning in robot manipulation tasks with sparse rewards from pixel observations.
Specifically, our technique involves using expert observations as intermediate visual goals for a goal-conditioned RL agent.
We demonstrate the efficacy of our method in five challenging block construction tasks in simulation and show that when combined with two state-of-the-art agents, our approach can significantly improve their performance.
arXiv Detail & Related papers (2023-06-24T05:57:05Z) - Pre-training Multi-task Contrastive Learning Models for Scientific
Literature Understanding [52.723297744257536]
Pre-trained language models (LMs) have shown effectiveness in scientific literature understanding tasks.
We propose a multi-task contrastive learning framework, SciMult, to facilitate common knowledge sharing across different literature understanding tasks.
arXiv Detail & Related papers (2023-05-23T16:47:22Z) - CLARE: Conservative Model-Based Reward Learning for Offline Inverse
Reinforcement Learning [26.05184273238923]
This work aims to tackle a major challenge in offline Inverse Reinforcement Learning (IRL)
We devise a principled algorithm (namely CLARE) that solves offline IRL efficiently via integrating "conservatism" into a learned reward function.
Our theoretical analysis provides an upper bound on the return gap between the learned policy and the expert policy.
arXiv Detail & Related papers (2023-02-09T17:16:29Z) - Reinforcement learning with Demonstrations from Mismatched Task under
Sparse Reward [7.51772160511614]
Reinforcement learning often suffer from the sparse reward issue in real-world robotics problems.
Prior works often assume that the learning agent and the expert aim to accomplish the same task, which requires collecting new data for every new task.
In this paper, we consider the case where the target task is mismatched from but similar with that of the expert.
Existing LfD methods can not effectively guide learning in mismatched new tasks with sparse rewards.
arXiv Detail & Related papers (2022-12-03T02:24:59Z) - Learning from Guided Play: A Scheduled Hierarchical Approach for
Improving Exploration in Adversarial Imitation Learning [7.51557557629519]
We present Learning from Guided Play (LfGP), a framework in which we leverage expert demonstrations of, in addition to a main task, multiple auxiliary tasks.
This affords many benefits: learning efficiency is improved for main tasks with challenging bottleneck transitions, expert data becomes reusable between tasks, and transfer learning through the reuse of learned auxiliary task models becomes possible.
arXiv Detail & Related papers (2021-12-16T14:58:08Z) - Learning Task Agnostic Skills with Data-driven Guidance [0.0]
This paper proposes a framework for guiding the skill discovery towards the subset of expert-visited states.
We apply our method in various reinforcement learning tasks and show that such a projection results in more useful behaviours.
arXiv Detail & Related papers (2021-08-04T06:53:10Z) - PEBBLE: Feedback-Efficient Interactive Reinforcement Learning via
Relabeling Experience and Unsupervised Pre-training [94.87393610927812]
We present an off-policy, interactive reinforcement learning algorithm that capitalizes on the strengths of both feedback and off-policy learning.
We demonstrate that our approach is capable of learning tasks of higher complexity than previously considered by human-in-the-loop methods.
arXiv Detail & Related papers (2021-06-09T14:10:50Z) - Bridging the Imitation Gap by Adaptive Insubordination [88.35564081175642]
We show that when the teaching agent makes decisions with access to privileged information, this information is marginalized during imitation learning.
We propose 'Adaptive Insubordination' (ADVISOR) to address this gap.
ADVISOR dynamically weights imitation and reward-based reinforcement learning losses during training, enabling on-the-fly switching between imitation and exploration.
arXiv Detail & Related papers (2020-07-23T17:59:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.