Intrinsic Motivation for Encouraging Synergistic Behavior
- URL: http://arxiv.org/abs/2002.05189v1
- Date: Wed, 12 Feb 2020 19:34:51 GMT
- Title: Intrinsic Motivation for Encouraging Synergistic Behavior
- Authors: Rohan Chitnis, Shubham Tulsiani, Saurabh Gupta, Abhinav Gupta
- Abstract summary: We study the role of intrinsic motivation as an exploration bias for reinforcement learning in sparse-reward synergistic tasks.
Our key idea is that a good guiding principle for intrinsic motivation in synergistic tasks is to take actions which affect the world in ways that would not be achieved if the agents were acting on their own.
- Score: 55.10275467562764
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the role of intrinsic motivation as an exploration bias for
reinforcement learning in sparse-reward synergistic tasks, which are tasks
where multiple agents must work together to achieve a goal they could not
individually. Our key idea is that a good guiding principle for intrinsic
motivation in synergistic tasks is to take actions which affect the world in
ways that would not be achieved if the agents were acting on their own. Thus,
we propose to incentivize agents to take (joint) actions whose effects cannot
be predicted via a composition of the predicted effect for each individual
agent. We study two instantiations of this idea, one based on the true states
encountered, and another based on a dynamics model trained concurrently with
the policy. While the former is simpler, the latter has the benefit of being
analytically differentiable with respect to the action taken. We validate our
approach in robotic bimanual manipulation and multi-agent locomotion tasks with
sparse rewards; we find that our approach yields more efficient learning than
both 1) training with only the sparse reward and 2) using the typical
surprise-based formulation of intrinsic motivation, which does not bias toward
synergistic behavior. Videos are available on the project webpage:
https://sites.google.com/view/iclr2020-synergistic.
Related papers
- DCIR: Dynamic Consistency Intrinsic Reward for Multi-Agent Reinforcement
Learning [84.22561239481901]
We propose a new approach that enables agents to learn whether their behaviors should be consistent with that of other agents.
We evaluate DCIR in multiple environments including Multi-agent Particle, Google Research Football and StarCraft II Micromanagement.
arXiv Detail & Related papers (2023-12-10T06:03:57Z) - Learning Goal-based Movement via Motivational-based Models in Cognitive
Mobile Robots [58.720142291102135]
Humans have needs motivating their behavior according to intensity and context.
We also create preferences associated with each action's perceived pleasure, which is susceptible to changes over time.
This makes decision-making more complex, requiring learning to balance needs and preferences according to the context.
arXiv Detail & Related papers (2023-02-20T04:52:24Z) - Collaborative Training of Heterogeneous Reinforcement Learning Agents in
Environments with Sparse Rewards: What and When to Share? [7.489793155793319]
This work focuses on combining information obtained through intrinsic motivation with the aim of having a more efficient exploration and faster learning.
Our results reveal different ways in which a collaborative framework with little additional computational cost can outperform an independent learning process without knowledge sharing.
arXiv Detail & Related papers (2022-02-24T16:15:51Z) - Contrastive Active Inference [12.361539023886161]
We propose a contrastive objective for active inference that reduces the computational burden in learning the agent's generative model and planning future actions.
Our method performs notably better than likelihood-based active inference in image-based tasks, while also being computationally cheaper and easier to train.
arXiv Detail & Related papers (2021-10-19T16:20:49Z) - Mutual Information State Intrinsic Control [91.38627985733068]
Intrinsically motivated RL attempts to remove this constraint by defining an intrinsic reward function.
Motivated by the self-consciousness concept in psychology, we make a natural assumption that the agent knows what constitutes itself.
We mathematically formalize this reward as the mutual information between the agent state and the surrounding state.
arXiv Detail & Related papers (2021-03-15T03:03:36Z) - Tracking Emotions: Intrinsic Motivation Grounded on Multi-Level
Prediction Error Dynamics [68.8204255655161]
We discuss how emotions arise when differences between expected and actual rates of progress towards a goal are experienced.
We present an intrinsic motivation architecture that generates behaviors towards self-generated and dynamic goals.
arXiv Detail & Related papers (2020-07-29T06:53:13Z) - Learning intuitive physics and one-shot imitation using
state-action-prediction self-organizing maps [0.0]
Humans learn by exploration and imitation, build causal models of the world, and use both to flexibly solve new tasks.
We suggest a simple but effective unsupervised model which develops such characteristics.
We demonstrate its performance on a set of several related, but different one-shot imitation tasks, which the agent flexibly solves in an active inference style.
arXiv Detail & Related papers (2020-07-03T12:29:11Z) - Mutual Information-based State-Control for Intrinsically Motivated
Reinforcement Learning [102.05692309417047]
In reinforcement learning, an agent learns to reach a set of goals by means of an external reward signal.
In the natural world, intelligent organisms learn from internal drives, bypassing the need for external signals.
We propose to formulate an intrinsic objective as the mutual information between the goal states and the controllable states.
arXiv Detail & Related papers (2020-02-05T19:21:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.