Coagent Networks: Generalized and Scaled
- URL: http://arxiv.org/abs/2305.09838v1
- Date: Tue, 16 May 2023 22:41:56 GMT
- Title: Coagent Networks: Generalized and Scaled
- Authors: James E. Kostas, Scott M. Jordan, Yash Chandak, Georgios Theocharous,
Dhawal Gupta, Martha White, Bruno Castro da Silva, Philip S. Thomas
- Abstract summary: Coagent networks for reinforcement learning (RL) provide a powerful and flexible framework for deriving principled learning rules.
This work generalizes the coagent theory and learning rules provided by previous works.
We show that a coagent algorithm with a policy network that does not use backpropagation can scale to a challenging RL domain.
- Score: 44.06183176712763
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Coagent networks for reinforcement learning (RL) [Thomas and Barto, 2011]
provide a powerful and flexible framework for deriving principled learning
rules for arbitrary stochastic neural networks. The coagent framework offers an
alternative to backpropagation-based deep learning (BDL) that overcomes some of
backpropagation's main limitations. For example, coagent networks can compute
different parts of the network \emph{asynchronously} (at different rates or at
different times), can incorporate non-differentiable components that cannot be
used with backpropagation, and can explore at levels higher than their action
spaces (that is, they can be designed as hierarchical networks for exploration
and/or temporal abstraction). However, the coagent framework is not just an
alternative to BDL; the two approaches can be blended: BDL can be combined with
coagent learning rules to create architectures with the advantages of both
approaches. This work generalizes the coagent theory and learning rules
provided by previous works; this generalization provides more flexibility for
network architecture design within the coagent framework. This work also
studies one of the chief disadvantages of coagent networks: high variance
updates for networks that have many coagents and do not use backpropagation. We
show that a coagent algorithm with a policy network that does not use
backpropagation can scale to a challenging RL domain with a high-dimensional
state and action space (the MuJoCo Ant environment), learning reasonable
(although not state-of-the-art) policies. These contributions motivate and
provide a more general theoretical foundation for future work that studies
coagent networks.
Related papers
- Block-Operations: Using Modular Routing to Improve Compositional Generalization [17.525220958618988]
We introduce the Multiplexer, a new architectural component that enhances the Feed Forward Neural Network (FNN)
We experimentally confirm that Multiplexers exhibit strong compositional generalization.
arXiv Detail & Related papers (2024-08-01T12:28:22Z) - Inter-Cell Network Slicing With Transfer Learning Empowered Multi-Agent
Deep Reinforcement Learning [6.523367518762879]
Network slicing enables operators to efficiently support diverse applications on a common physical infrastructure.
The ever-increasing densification of network deployment leads to complex and non-trivial inter-cell interference.
We develop a DIRP algorithm with multiple deep reinforcement learning (DRL) agents to cooperatively optimize resource partition in individual cells.
arXiv Detail & Related papers (2023-06-20T14:14:59Z) - Diversity Through Exclusion (DTE): Niche Identification for
Reinforcement Learning through Value-Decomposition [63.67574523750839]
We propose a generic reinforcement learning (RL) algorithm that performs better than baseline deep Q-learning algorithms in environments with multiple variably-valued niches.
We show that agents trained this way can escape poor-but-attractive local optima to instead converge to harder-to-discover higher value strategies.
arXiv Detail & Related papers (2023-02-02T16:00:19Z) - Coordinated Reinforcement Learning for Optimizing Mobile Networks [6.924083445159127]
We show how to use coordination graphs and reinforcement learning in a complex application involving hundreds of cooperating agents.
We show empirically that coordinated reinforcement learning outperforms other methods.
arXiv Detail & Related papers (2021-09-30T14:46:18Z) - Phase Retrieval using Expectation Consistent Signal Recovery Algorithm
based on Hypernetwork [73.94896986868146]
Phase retrieval is an important component in modern computational imaging systems.
Recent advances in deep learning have opened up a new possibility for robust and fast PR.
We develop a novel framework for deep unfolding to overcome the existing limitations.
arXiv Detail & Related papers (2021-01-12T08:36:23Z) - F2A2: Flexible Fully-decentralized Approximate Actor-critic for
Cooperative Multi-agent Reinforcement Learning [110.35516334788687]
Decentralized multi-agent reinforcement learning algorithms are sometimes unpractical in complicated applications.
We propose a flexible fully decentralized actor-critic MARL framework, which can handle large-scale general cooperative multi-agent setting.
Our framework can achieve scalability and stability for large-scale environment and reduce information transmission.
arXiv Detail & Related papers (2020-04-17T14:56:29Z) - Dynamic Hierarchical Mimicking Towards Consistent Optimization
Objectives [73.15276998621582]
We propose a generic feature learning mechanism to advance CNN training with enhanced generalization ability.
Partially inspired by DSN, we fork delicately designed side branches from the intermediate layers of a given neural network.
Experiments on both category and instance recognition tasks demonstrate the substantial improvements of our proposed method.
arXiv Detail & Related papers (2020-03-24T09:56:13Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z) - Coagent Networks Revisited [10.45819881530349]
Coagent networks formalize the concept of arbitrary networks of agents that collaborate to take actions in a reinforcement learning environment.
We first provide a unifying perspective on the many diverse examples that fall under coagent networks.
We do so by formalizing the rules of execution in a coagent network, enabled by the novel and intuitive idea of execution paths.
arXiv Detail & Related papers (2020-01-28T17:31:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.