Evaluating Team Skill Aggregation in Online Competitive Games
- URL: http://arxiv.org/abs/2106.11397v1
- Date: Mon, 21 Jun 2021 20:17:36 GMT
- Title: Evaluating Team Skill Aggregation in Online Competitive Games
- Authors: Arman Dehpanah, Muheeb Faizan Ghori, Jonathan Gemmell, Bamshad
Mobasher
- Abstract summary: We present an analysis of the impact of two new aggregation methods on the predictive performance of rating systems.
Our evaluations show the superiority of the MAX method over the other two methods in the majority of the tested cases.
Results of this study highlight the necessity of devising more elaborated methods for calculating a team's performance.
- Score: 4.168733556014873
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One of the main goals of online competitive games is increasing player
engagement by ensuring fair matches. These games use rating systems for
creating balanced match-ups. Rating systems leverage statistical estimation to
rate players' skills and use skill ratings to predict rank before matching
players. Skill ratings of individual players can be aggregated to compute the
skill level of a team. While research often aims to improve the accuracy of
skill estimation and fairness of match-ups, less attention has been given to
how the skill level of a team is calculated from the skill level of its
members. In this paper, we propose two new aggregation methods and compare them
with a standard approach extensively used in the research literature. We
present an exhaustive analysis of the impact of these methods on the predictive
performance of rating systems. We perform our experiments using three popular
rating systems, Elo, Glicko, and TrueSkill, on three real-world datasets
including over 100,000 battle royale and head-to-head matches. Our evaluations
show the superiority of the MAX method over the other two methods in the
majority of the tested cases, implying that the overall performance of a team
is best determined by the performance of its most skilled member. The results
of this study highlight the necessity of devising more elaborated methods for
calculating a team's performance -- methods covering different aspects of
players' behavior such as skills, strategy, or goals.
Related papers
- Skill Issues: An Analysis of CS:GO Skill Rating Systems [0.24578723416255746]
Elo, Glicko2 and TrueSkill are studied through the lens of surrogate modelling.
We look at both overall performance and data efficiency, and perform a sensitivity analysis based on a large dataset of Counter-Strike: Global Offensive matches.
arXiv Detail & Related papers (2024-10-01T23:19:31Z) - ApproxED: Approximate exploitability descent via learned best responses [61.17702187957206]
We study the problem of finding an approximate Nash equilibrium of games with continuous action sets.
We propose two new methods that minimize an approximation of exploitability with respect to the strategy profile.
arXiv Detail & Related papers (2023-01-20T23:55:30Z) - A Tale of HodgeRank and Spectral Method: Target Attack Against Rank
Aggregation Is the Fixed Point of Adversarial Game [153.74942025516853]
The intrinsic vulnerability of the rank aggregation methods is not well studied in the literature.
In this paper, we focus on the purposeful adversary who desires to designate the aggregated results by modifying the pairwise data.
The effectiveness of the suggested target attack strategies is demonstrated by a series of toy simulations and several real-world data experiments.
arXiv Detail & Related papers (2022-09-13T05:59:02Z) - QuickSkill: Novice Skill Estimation in Online Multiplayer Games [19.364132825629465]
Current matchmaking rating algorithms require considerable amount of games for learning the true skill of a new player.
This is known as the ''cold-start'' problem for matchmaking rating algorithms.
This paper proposes QuickSKill, a deep learning based novice skill estimation framework.
arXiv Detail & Related papers (2022-08-15T11:59:05Z) - Behavioral Player Rating in Competitive Online Shooter Games [3.203973145772361]
In this paper, we engineer several features from in-game statistics to model players and create ratings that accurately represent their behavior and true performance level.
Our results show that the behavioral ratings present more accurate performance estimations while maintaining the interpretability of the created representations.
Considering different aspects of the playing behavior of players and using behavioral ratings for matchmaking can lead to match-ups that are more aligned with players' goals and interests.
arXiv Detail & Related papers (2022-07-01T16:23:01Z) - Collusion Detection in Team-Based Multiplayer Games [57.153233321515984]
We propose a system that detects colluding behaviors in team-based multiplayer games.
The proposed method analyzes the players' social relationships paired with their in-game behavioral patterns.
We then automate the detection using Isolation Forest, an unsupervised learning technique specialized in highlighting outliers.
arXiv Detail & Related papers (2022-03-10T02:37:39Z) - The Evaluation of Rating Systems in Team-based Battle Royale Games [4.168733556014873]
This paper explores the utility of several metrics for evaluating three popular rating systems on a real-world dataset of over 25,000 team battle royale matches.
normalized discounted cumulative gain (NDCG) demonstrated more reliable performance and more flexibility.
arXiv Detail & Related papers (2021-05-28T19:22:07Z) - An Empirical Study on the Generalization Power of Neural Representations
Learned via Visual Guessing Games [79.23847247132345]
This work investigates how well an artificial agent can benefit from playing guessing games when later asked to perform on novel NLP downstream tasks such as Visual Question Answering (VQA)
We propose two ways to exploit playing guessing games: 1) a supervised learning scenario in which the agent learns to mimic successful guessing games and 2) a novel way for an agent to play by itself, called Self-play via Iterated Experience Learning (SPIEL)
arXiv Detail & Related papers (2021-01-31T10:30:48Z) - Faster Algorithms for Optimal Ex-Ante Coordinated Collusive Strategies
in Extensive-Form Zero-Sum Games [123.76716667704625]
We focus on the problem of finding an optimal strategy for a team of two players that faces an opponent in an imperfect-information zero-sum extensive-form game.
In that setting, it is known that the best the team can do is sample a profile of potentially randomized strategies (one per player) from a joint (a.k.a. correlated) probability distribution at the beginning of the game.
We provide an algorithm that computes such an optimal distribution by only using profiles where only one of the team members gets to randomize in each profile.
arXiv Detail & Related papers (2020-09-21T17:51:57Z) - Competitive Balance in Team Sports Games [8.321949054700086]
We show that using final score difference provides yet a better prediction metric for competitive balance.
We also show that a linear model trained on a carefully selected set of team and individual features achieves almost the performance of the more powerful neural network model.
arXiv Detail & Related papers (2020-06-24T14:19:07Z) - Learning from Learners: Adapting Reinforcement Learning Agents to be
Competitive in a Card Game [71.24825724518847]
We present a study on how popular reinforcement learning algorithms can be adapted to learn and to play a real-world implementation of a competitive multiplayer card game.
We propose specific training and validation routines for the learning agents, in order to evaluate how the agents learn to be competitive and explain how they adapt to each others' playing style.
arXiv Detail & Related papers (2020-04-08T14:11:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.