Fair Shares: Feasibility, Domination and Incentives
- URL: http://arxiv.org/abs/2205.07519v1
- Date: Mon, 16 May 2022 08:52:42 GMT
- Title: Fair Shares: Feasibility, Domination and Incentives
- Authors: Moshe Babaioff and Uriel Feige
- Abstract summary: We consider fair allocation of a set $M$ of indivisible goods to $n$ equally-entitled agents, with no monetary transfers.
For additive valuations we present shares that are feasible, self-maximizing and time computable.
- Score: 6.878547831852427
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider fair allocation of a set $M$ of indivisible goods to $n$
equally-entitled agents, with no monetary transfers. Every agent $i$ has a
valuation $v_i$ from some given class of valuation functions. A share $s$ is a
function that maps a pair $(v_i,n)$ to a value, with the interpretation that if
an allocation of $M$ to $n$ agents fails to give agent $i$ a bundle of value at
least equal to $s(v_i,n)$, this serves as evidence that the allocation is not
fair towards $i$. For such an interpretation to make sense, we would like the
share to be feasible, meaning that for any valuations in the class, there is an
allocation that gives every agent at least her share. The maximin share was a
natural candidate for a feasible share for additive valuations. However,
Kurokawa, Procaccia and Wang [2018] show that it is not feasible.
We initiate a systematic study of the family of feasible shares. We say that
a share is \emph{self maximizing} if truth-telling maximizes the implied
guarantee. We show that every feasible share is dominated by some
self-maximizing and feasible share. We seek to identify those self-maximizing
feasible shares that are polynomial time computable, and offer the highest
share values. We show that a SM-dominating feasible share -- one that dominates
every self-maximizing (SM) feasible share -- does not exist for additive
valuations (and beyond). Consequently, we relax the domination property to that
of domination up to a multiplicative factor of $\rho$ (called
$\rho$-dominating). For additive valuations we present shares that are
feasible, self-maximizing and polynomial-time computable. For $n$ agents we
present such a share that is $\frac{2n}{3n-1}$-dominating. For two agents we
present such a share that is $(1 - \epsilon)$-dominating. Moreover, for these
shares we present poly-time algorithms that compute allocations that give every
agent at least her share.
Related papers
- Mind the Gap: A Causal Perspective on Bias Amplification in Prediction & Decision-Making [58.06306331390586]
We introduce the notion of a margin complement, which measures how much a prediction score $S$ changes due to a thresholding operation.
We show that under suitable causal assumptions, the influences of $X$ on the prediction score $S$ are equal to the influences of $X$ on the true outcome $Y$.
arXiv Detail & Related papers (2024-05-24T11:22:19Z) - Near-Minimax-Optimal Risk-Sensitive Reinforcement Learning with CVaR [58.40575099910538]
We study risk-sensitive Reinforcement Learning (RL), focusing on the objective of Conditional Value at Risk (CVaR) with risk tolerance $tau$.
We show the minimax CVaR regret rate is $Omega(sqrttau-1AK)$, where $A$ is the number of actions and $K$ is the number of episodes.
We show that our algorithm achieves the optimal regret of $widetilde O(tau-1sqrtSAK)$ under a continuity assumption and in general attains a near
arXiv Detail & Related papers (2023-02-07T02:22:31Z) - Dividing Good and Better Items Among Agents with Bivalued Submodular
Valuations [20.774185319381985]
We study the problem of fairly allocating a set of indivisible goods among agents with bivalued submodular valuations.
We show that neither the leximin nor the MNW allocation is guaranteed to be envy free up to one good.
arXiv Detail & Related papers (2023-02-06T19:41:28Z) - Minimax-Optimal Multi-Agent RL in Zero-Sum Markov Games With a
Generative Model [50.38446482252857]
Two-player zero-sum Markov games are arguably the most basic setting in multi-agent reinforcement learning.
We develop a learning algorithm that learns an $varepsilon$-approximate Markov NE policy using $$ widetildeObigg.
We derive a refined regret bound for FTRL that makes explicit the role of variance-type quantities.
arXiv Detail & Related papers (2022-08-22T17:24:55Z) - Collaborative Multi-agent Stochastic Linear Bandits [28.268809091816287]
We study a collaborative multi-agent linear bandit setting, where $N$ agents that form a network communicate locally to minimize their overall regret.
All the agents observe the corresponding rewards of the played actions and use an accelerated consensus procedure to compute an estimate of the average of the rewards obtained by all the agents.
arXiv Detail & Related papers (2022-05-12T19:46:35Z) - Nearly Minimax Algorithms for Linear Bandits with Shared Representation [86.79657561369397]
We consider the setting where we play $M$ linear bandits with dimension $d$, each for $T$ rounds, and these $M$ bandit tasks share a common $k(ll d)$ dimensional linear representation.
We come up with novel algorithms that achieve $widetildeOleft(dsqrtkMT + kMsqrtTright)$ regret bounds, which matches the known minimax regret lower bound up to logarithmic factors.
arXiv Detail & Related papers (2022-03-29T15:27:13Z) - (Almost) Envy-Free, Proportional and Efficient Allocations of an
Indivisible Mixed Manna [10.933894827834825]
We study the problem of finding fair and efficient allocations of a set of indivisible items to a set of agents.
As fairness notions, we consider arguably the strongest possible relaxations of envy-freeness and proportionality.
arXiv Detail & Related papers (2022-02-06T01:29:50Z) - Coordinated Attacks against Contextual Bandits: Fundamental Limits and
Defense Mechanisms [75.17357040707347]
Motivated by online recommendation systems, we propose the problem of finding the optimal policy in contextual bandits.
The goal is to robustly learn the policy that maximizes rewards for good users with as few user interactions as possible.
We show we can achieve an $tildeO(min(S,A)cdot alpha/epsilon2)$ upper-bound, by employing efficient robust mean estimators.
arXiv Detail & Related papers (2022-01-30T01:45:13Z) - Fine-Grained Gap-Dependent Bounds for Tabular MDPs via Adaptive
Multi-Step Bootstrap [84.66885506098724]
This paper presents a new model-free algorithm for episodic finite-horizon Markov Decision Processes (MDP), Adaptive Multi-step Bootstrap (AMB)
We show AMB achieves a gap-dependent regret bound that only scales with the sum of the inverse of the sub-optimality gaps.
We also show AMB suffers an additional $frac|Z_mul|Delta_min$ regret, where $Z_mul$ is the set of state-action pairs $(s,a)$'s satisfying $a$ is a non-unique optimal action for
arXiv Detail & Related papers (2021-02-09T07:46:34Z) - Finite Continuum-Armed Bandits [0.0]
We consider a situation where an agent has $T$ ressources to be allocated to a larger number of actions.
The goal of the agent is to maximize her cumulative reward.
arXiv Detail & Related papers (2020-10-23T08:48:45Z) - A No-Free-Lunch Theorem for MultiTask Learning [19.645741778058227]
We consider a seemingly favorable classification scenario where all tasks $P_t$ share a common optimal classifier $h*,$.
We show that, even though such regimes admit minimax rates accounting for both $n$ and $N$, no adaptive algorithm exists.
arXiv Detail & Related papers (2020-06-29T03:03:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.