Adversarial Intrinsic Motivation for Reinforcement Learning
- URL: http://arxiv.org/abs/2105.13345v2
- Date: Sun, 30 May 2021 22:13:00 GMT
- Title: Adversarial Intrinsic Motivation for Reinforcement Learning
- Authors: Ishan Durugkar, Mauricio Tec, Scott Niekum, Peter Stone
- Abstract summary: We investigate whether the Wasserstein-1 distance between a policy's state visitation distribution and a target distribution can be utilized effectively for reinforcement learning tasks.
Our approach, termed Adversarial Intrinsic Motivation (AIM), estimates this Wasserstein-1 distance through its dual objective and uses it to compute a supplemental reward function.
- Score: 60.322878138199364
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning with an objective to minimize the mismatch with a reference
distribution has been shown to be useful for generative modeling and imitation
learning. In this paper, we investigate whether one such objective, the
Wasserstein-1 distance between a policy's state visitation distribution and a
target distribution, can be utilized effectively for reinforcement learning
(RL) tasks. Specifically, this paper focuses on goal-conditioned reinforcement
learning where the idealized (unachievable) target distribution has full
measure at the goal. We introduce a quasimetric specific to Markov Decision
Processes (MDPs), and show that the policy that minimizes the Wasserstein-1
distance of its state visitation distribution to this target distribution under
this quasimetric is the policy that reaches the goal in as few steps as
possible. Our approach, termed Adversarial Intrinsic Motivation (AIM),
estimates this Wasserstein-1 distance through its dual objective and uses it to
compute a supplemental reward function. Our experiments show that this reward
function changes smoothly with respect to transitions in the MDP and assists
the agent in learning. Additionally, we combine AIM with Hindsight Experience
Replay (HER) and show that the resulting algorithm accelerates learning
significantly on several simulated robotics tasks when compared to HER with a
sparse positive reward at the goal state.
Related papers
- A Unifying Framework for Action-Conditional Self-Predictive Reinforcement Learning [48.59516337905877]
Learning a good representation is a crucial challenge for Reinforcement Learning (RL) agents.
Recent work has developed theoretical insights into these algorithms.
We take a step towards bridging the gap between theory and practice by analyzing an action-conditional self-predictive objective.
arXiv Detail & Related papers (2024-06-04T07:22:12Z) - Discrete Factorial Representations as an Abstraction for Goal
Conditioned Reinforcement Learning [99.38163119531745]
We show that applying a discretizing bottleneck can improve performance in goal-conditioned RL setups.
We experimentally prove the expected return on out-of-distribution goals, while still allowing for specifying goals with expressive structure.
arXiv Detail & Related papers (2022-11-01T03:31:43Z) - TarGF: Learning Target Gradient Field for Object Rearrangement [8.49306925839127]
We focus on a more practical setting in object rearrangement, i.e., rearranging objects from shuffled layouts to a normative target distribution.
It is hard to describe the target distribution (goal specification) for reward engineering or collect expert trajectories as demonstrations.
We employ the score-matching objective to train a Target Gradient Field (TarGF), indicating a direction on each object to increase the likelihood of the target distribution.
arXiv Detail & Related papers (2022-09-02T07:20:34Z) - Goal-Conditioned Q-Learning as Knowledge Distillation [136.79415677706612]
We explore a connection between off-policy reinforcement learning in goal-conditioned settings and knowledge distillation.
We empirically show that this can improve the performance of goal-conditioned off-policy reinforcement learning when the space of goals is high-dimensional.
We also show that this technique can be adapted to allow for efficient learning in the case of multiple simultaneous sparse goals.
arXiv Detail & Related papers (2022-08-28T22:01:10Z) - Generative multitask learning mitigates target-causing confounding [61.21582323566118]
We propose a simple and scalable approach to causal representation learning for multitask learning.
The improvement comes from mitigating unobserved confounders that cause the targets, but not the input.
Our results on the Attributes of People and Taskonomy datasets reflect the conceptual improvement in robustness to prior probability shift.
arXiv Detail & Related papers (2022-02-08T20:42:14Z) - Goal-Aware Cross-Entropy for Multi-Target Reinforcement Learning [15.33496710690063]
We propose goal-aware cross-entropy (GACE) loss, that can be utilized in a self-supervised way.
We then devise goal-discriminative attention networks (GDAN) which utilize the goal-relevant information to focus on the given instruction.
arXiv Detail & Related papers (2021-10-25T14:24:39Z) - Physical Reasoning Using Dynamics-Aware Models [32.402950370430496]
This study aims to address the limitation by augmenting the reward value with additional supervisory signals about object dynamics.
Specifically,we define a distance measure between the trajectory of two target objects, and use this distance measure to characterize the similarity of two environment rollouts.
We train the model to correctly rank rollouts according to this measure in addition to predicting the correct reward.
arXiv Detail & Related papers (2021-02-20T12:56:16Z) - Automatic Curriculum Learning through Value Disagreement [95.19299356298876]
Continually solving new, unsolved tasks is the key to learning diverse behaviors.
In the multi-task domain, where an agent needs to reach multiple goals, the choice of training goals can largely affect sample efficiency.
We propose setting up an automatic curriculum for goals that the agent needs to solve.
We evaluate our method across 13 multi-goal robotic tasks and 5 navigation tasks, and demonstrate performance gains over current state-of-the-art methods.
arXiv Detail & Related papers (2020-06-17T03:58:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.