PeerNomination: Relaxing Exactness for Increased Accuracy in Peer
Selection
- URL: http://arxiv.org/abs/2004.14939v1
- Date: Thu, 30 Apr 2020 16:39:47 GMT
- Title: PeerNomination: Relaxing Exactness for Increased Accuracy in Peer
Selection
- Authors: Nicholas Mattei, Paolo Turrini, Stanislav Zhydkov
- Abstract summary: In peer selection agents must choose a subset of themselves for an award or a prize.
As agents are self-interested, we want to design algorithms that are impartial, so that an individual agent cannot affect their own chance of being selected.
Here, we present a novel algorithm for impartial peer selection, PeerNomination, and provide a theoretical analysis of its accuracy.
- Score: 30.779586758074206
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In peer selection agents must choose a subset of themselves for an award or a
prize. As agents are self-interested, we want to design algorithms that are
impartial, so that an individual agent cannot affect their own chance of being
selected. This problem has broad application in resource allocation and
mechanism design and has received substantial attention in the artificial
intelligence literature. Here, we present a novel algorithm for impartial peer
selection, PeerNomination, and provide a theoretical analysis of its accuracy.
Our algorithm possesses various desirable features. In particular, it does not
require an explicit partitioning of the agents, as previous algorithms in the
literature. We show empirically that it achieves higher accuracy than the
exiting algorithms over several metrics.
Related papers
- Multi-Armed Bandits with Abstention [62.749500564313834]
We introduce a novel extension of the canonical multi-armed bandit problem that incorporates an additional strategic element: abstention.
In this enhanced framework, the agent is not only tasked with selecting an arm at each time step, but also has the option to abstain from accepting the instantaneous reward before observing it.
arXiv Detail & Related papers (2024-02-23T06:27:12Z) - Scalable Decentralized Algorithms for Online Personalized Mean Estimation [12.002609934938224]
This study focuses on a simplified version of the overarching problem, where each agent collects samples from a real-valued distribution over time to estimate its mean.
We introduce two collaborative mean estimation algorithms: one draws inspiration from belief propagation, while the other employs a consensus-based approach.
arXiv Detail & Related papers (2024-02-20T08:30:46Z) - A Gold Standard Dataset for the Reviewer Assignment Problem [117.59690218507565]
"Similarity score" is a numerical estimate of the expertise of a reviewer in reviewing a paper.
Our dataset consists of 477 self-reported expertise scores provided by 58 researchers.
For the task of ordering two papers in terms of their relevance for a reviewer, the error rates range from 12%-30% in easy cases to 36%-43% in hard cases.
arXiv Detail & Related papers (2023-03-23T16:15:03Z) - Stochastic Differentially Private and Fair Learning [7.971065005161566]
We provide the first differentially private algorithm for fair learning that is guaranteed to converge.
Our framework is flexible enough to permit different fairness, including demographic parity and equalized odds.
Our algorithm can be applied to non-binary classification tasks with multiple (non-binary) sensitive attributes.
arXiv Detail & Related papers (2022-10-17T06:54:57Z) - Multi-Agent Algorithmic Recourse [7.23389716633927]
We show that when the assumption of a single agent environment is relaxed, current approaches to algorithmic recourse fail to guarantee certain ethically desirable properties.
We propose a new game theory inspired framework for providing algorithmic recourse in a multi-agent environment.
arXiv Detail & Related papers (2021-10-01T22:54:47Z) - Machine Learning for Online Algorithm Selection under Censored Feedback [71.6879432974126]
In online algorithm selection (OAS), instances of an algorithmic problem class are presented to an agent one after another, and the agent has to quickly select a presumably best algorithm from a fixed set of candidate algorithms.
For decision problems such as satisfiability (SAT), quality typically refers to the algorithm's runtime.
In this work, we revisit multi-armed bandit algorithms for OAS and discuss their capability of dealing with the problem.
We adapt them towards runtime-oriented losses, allowing for partially censored data while keeping a space- and time-complexity independent of the time horizon.
arXiv Detail & Related papers (2021-09-13T18:10:52Z) - Selection-Expansion: A Unifying Framework for Motion-Planning and
Diversity Search Algorithms [69.87173070473717]
We investigate the properties of two diversity search algorithms, the Novelty Search and the Goal Exploration Process algorithms.
The relation to MP algorithms reveals that the smoothness, or lack of smoothness of the mapping between the policy parameter space and the outcome space plays a key role in the search efficiency.
arXiv Detail & Related papers (2021-04-10T13:52:27Z) - Double Coverage with Machine-Learned Advice [100.23487145400833]
We study the fundamental online $k$-server problem in a learning-augmented setting.
We show that our algorithm achieves for any k an almost optimal consistency-robustness tradeoff.
arXiv Detail & Related papers (2021-03-02T11:04:33Z) - Generalization in portfolio-based algorithm selection [97.74604695303285]
We provide the first provable guarantees for portfolio-based algorithm selection.
We show that if the portfolio is large, overfitting is inevitable, even with an extremely simple algorithm selector.
arXiv Detail & Related papers (2020-12-24T16:33:17Z) - Learning and Strongly Truthful Multi-Task Peer Prediction: A Variational
Approach [8.932080210400535]
We design a family of mechanisms with a scoring function that maps a pair of reports to a score.
We show how to derive good bounds on the number of tasks required for different types of priors.
This is the first peer-prediction mechanism on continuous signals designed for the multi-task setting.
arXiv Detail & Related papers (2020-09-30T15:09:56Z) - No-Regret and Incentive-Compatible Online Learning [29.267666165169324]
We study online learning settings in which experts act strategically to maximize their influence on the learning algorithm's predictions.
We want the learning algorithm to be no-regret with respect to the best fixed expert in hindsight.
We provide algorithms that achieve no regret and incentive compatibility for experts for both the full and partial information settings.
arXiv Detail & Related papers (2020-02-20T16:21:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.