Evaluating and Rewarding Teamwork Using Cooperative Game Abstractions
- URL: http://arxiv.org/abs/2006.09538v1
- Date: Tue, 16 Jun 2020 22:03:36 GMT
- Title: Evaluating and Rewarding Teamwork Using Cooperative Game Abstractions
- Authors: Tom Yan, Christian Kroer, Alexander Peysakhovich
- Abstract summary: We use cooperative game theory to study teams of artificial RL agents as well as real world teams from professional sports.
We introduce a parametric model called cooperative game abstractions (CGAs) for estimating CFs from data.
We provide identification results and sample bounds complexity for CGA models as well as error bounds in the estimation of the Shapley Value using CGAs.
- Score: 103.3630903577951
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Can we predict how well a team of individuals will perform together? How
should individuals be rewarded for their contributions to the team performance?
Cooperative game theory gives us a powerful set of tools for answering these
questions: the Characteristic Function (CF) and solution concepts like the
Shapley Value (SV). There are two major difficulties in applying these
techniques to real world problems: first, the CF is rarely given to us and
needs to be learned from data. Second, the SV is combinatorial in nature. We
introduce a parametric model called cooperative game abstractions (CGAs) for
estimating CFs from data. CGAs are easy to learn, readily interpretable, and
crucially allow linear-time computation of the SV. We provide identification
results and sample complexity bounds for CGA models as well as error bounds in
the estimation of the SV using CGAs. We apply our methods to study teams of
artificial RL agents as well as real world teams from professional sports.
Related papers
- Open Ad Hoc Teamwork with Cooperative Game Theory [28.605478081031215]
Ad hoc teamwork poses a challenging problem, requiring the design of an agent to collaborate with teammates without prior coordination or joint training.
One promising solution is leveraging the generalizability of graph neural networks to handle an unrestricted number of agents.
We propose a novel algorithm named CIAO, based on the game's framework, with additional provable implementation tricks that can facilitate learning.
arXiv Detail & Related papers (2024-02-23T11:04:33Z) - SMaRt: Improving GANs with Score Matching Regularity [94.81046452865583]
Generative adversarial networks (GANs) usually struggle in learning from highly diverse data, whose underlying manifold is complex.
We show that score matching serves as a promising solution to this issue thanks to its capability of persistently pushing the generated data points towards the real data manifold.
We propose to improve the optimization of GANs with score matching regularity (SMaRt)
arXiv Detail & Related papers (2023-11-30T03:05:14Z) - A General Learning Framework for Open Ad Hoc Teamwork Using Graph-based
Policy Learning [11.998708550268978]
We develop a class of solutions for open ad hoc teamwork under full and partial observability.
We show that our solution can learn efficient policies in open ad hoc teamwork in fully and partially observable cases.
arXiv Detail & Related papers (2022-10-11T13:44:44Z) - Neural Payoff Machines: Predicting Fair and Stable Payoff Allocations
Among Team Members [13.643650155415484]
We show how cooperative game-theoretic solutions can be distilled into a learned model by training neural networks.
Our approach creates models that can generalize to games far from the training distribution.
An important application of our framework is Explainable AI.
arXiv Detail & Related papers (2022-08-18T12:33:09Z) - Towards convergence to Nash equilibria in two-team zero-sum games [17.4461045395989]
Two-team zero-sum games are defined as multi-player games where players are split into two competing sets of agents.
We focus on the solution concept of Nash equilibria (NE)
We show that computing NE for this class of games is $textithard$ for the complexity class $mathrm$.
arXiv Detail & Related papers (2021-11-07T21:15:35Z) - Multi-task Learning of Order-Consistent Causal Graphs [59.9575145128345]
We consider the problem of discovering $K related Gaussian acyclic graphs (DAGs)
Under multi-task learning setting, we propose a $l_1/l$-regularized maximum likelihood estimator (MLE) for learning $K$ linear structural equation models.
We theoretically show that the joint estimator, by leveraging data across related tasks, can achieve a better sample complexity for recovering the causal order.
arXiv Detail & Related papers (2021-11-03T22:10:18Z) - An Empirical Study on the Generalization Power of Neural Representations
Learned via Visual Guessing Games [79.23847247132345]
This work investigates how well an artificial agent can benefit from playing guessing games when later asked to perform on novel NLP downstream tasks such as Visual Question Answering (VQA)
We propose two ways to exploit playing guessing games: 1) a supervised learning scenario in which the agent learns to mimic successful guessing games and 2) a novel way for an agent to play by itself, called Self-play via Iterated Experience Learning (SPIEL)
arXiv Detail & Related papers (2021-01-31T10:30:48Z) - Faster Algorithms for Optimal Ex-Ante Coordinated Collusive Strategies
in Extensive-Form Zero-Sum Games [123.76716667704625]
We focus on the problem of finding an optimal strategy for a team of two players that faces an opponent in an imperfect-information zero-sum extensive-form game.
In that setting, it is known that the best the team can do is sample a profile of potentially randomized strategies (one per player) from a joint (a.k.a. correlated) probability distribution at the beginning of the game.
We provide an algorithm that computes such an optimal distribution by only using profiles where only one of the team members gets to randomize in each profile.
arXiv Detail & Related papers (2020-09-21T17:51:57Z) - Finding Core Members of Cooperative Games using Agent-Based Modeling [0.0]
Agent-based modeling (ABM) is a powerful paradigm to gain insight into social phenomena.
In this paper, a algorithm is developed that can be embedded into an ABM to allow the agents to find coalition.
arXiv Detail & Related papers (2020-08-30T17:38:43Z) - Unshuffling Data for Improved Generalization [65.57124325257409]
Generalization beyond the training distribution is a core challenge in machine learning.
We show that partitioning the data into well-chosen, non-i.i.d. subsets treated as multiple training environments can guide the learning of models with better out-of-distribution generalization.
arXiv Detail & Related papers (2020-02-27T03:07:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.