Personalized Help for Optimizing Low-Skilled Users' Strategy
- URL: http://arxiv.org/abs/2411.09109v2
- Date: Mon, 17 Feb 2025 23:36:13 GMT
- Title: Personalized Help for Optimizing Low-Skilled Users' Strategy
- Authors: Feng Gu, Wichayaporn Wongkamjan, Jonathan K. Kummerfeld, Denis Peskoff, Jonathan May, Jordan Boyd-Graber,
- Abstract summary: We augment CICERO, a natural language agent, to generate both move and message advice based on player intentions.
A dozen Diplomacy games with novice and experienced players, with varying advice settings, show that some of the generated advice is beneficial.
- Score: 36.00722599200535
- License:
- Abstract: AIs can beat humans in game environments; however, how helpful those agents are to human remains understudied. We augment CICERO, a natural language agent that demonstrates superhuman performance in Diplomacy, to generate both move and message advice based on player intentions. A dozen Diplomacy games with novice and experienced players, with varying advice settings, show that some of the generated advice is beneficial. It helps novices compete with experienced players and in some instances even surpass them. The mere presence of advice can be advantageous, even if players do not follow it.
Related papers
- Evaluating the Effects of AI Directors for Quest Selection [2.3941497253612085]
We focus on AI Directors, systems which can dynamically modify a game, that personalize the player experience to match the player's preference.
Our results show that a non-random AI Director provides a better player experience than a random AI Director.
arXiv Detail & Related papers (2024-09-30T18:16:38Z) - Reinforcement Learning for High-Level Strategic Control in Tower Defense Games [47.618236610219554]
In strategy games, one of the most important aspects of game design is maintaining a sense of challenge for players.
We propose an automated approach that combines traditional scripted methods with reinforcement learning.
Results show that combining a learned approach, such as reinforcement learning, with a scripted AI produces a higher-performing and more robust agent than using only AI.
arXiv Detail & Related papers (2024-06-12T08:06:31Z) - More Victories, Less Cooperation: Assessing Cicero's Diplomacy Play [28.004172388400132]
This work seeks to understand the degree to which Cicero succeeds at communication.
We run two dozen games with humans and Cicero, totaling over 200 human-player hours of competition.
While AI can consistently outplay human players, AI-Human communication is still limited because of AI's difficulty with deception and persuasion.
arXiv Detail & Related papers (2024-06-07T05:03:44Z) - DanZero+: Dominating the GuanDan Game through Reinforcement Learning [95.90682269990705]
We develop an AI program for an exceptionally complex and popular card game called GuanDan.
We first put forward an AI program named DanZero for this game.
In order to further enhance the AI's capabilities, we apply policy-based reinforcement learning algorithm to GuanDan.
arXiv Detail & Related papers (2023-12-05T08:07:32Z) - Adversarial Policies Beat Superhuman Go AIs [54.15639517188804]
We attack the state-of-the-art Go-playing AI system KataGo by training adversarial policies against it.
Our adversaries do not win by playing Go well. Instead, they trick KataGo into making serious blunders.
Our results demonstrate that even superhuman AI systems may harbor surprising failure modes.
arXiv Detail & Related papers (2022-11-01T03:13:20Z) - Mastering the Game of No-Press Diplomacy via Human-Regularized
Reinforcement Learning and Planning [95.78031053296513]
No-press Diplomacy is a complex strategy game involving both cooperation and competition.
We introduce a planning algorithm we call DiL-piKL that regularizes a reward-maximizing policy toward a human imitation-learned policy.
We show that DiL-piKL can be extended into a self-play reinforcement learning algorithm we call RL-DiL-piKL.
arXiv Detail & Related papers (2022-10-11T14:47:35Z) - Generative Personas That Behave and Experience Like Humans [3.611888922173257]
generative AI agents attempt to imitate particular playing behaviors represented as rules, rewards, or human demonstrations.
We extend the notion of behavioral procedural personas to cater for player experience, thus examining generative agents that can both behave and experience their game as humans would.
Our findings suggest that the generated agents exhibit distinctive play styles and experience responses of the human personas they were designed to imitate.
arXiv Detail & Related papers (2022-08-26T12:04:53Z) - The corruptive force of AI-generated advice [0.0]
We test whether AI-generated advice can corrupt people.
We also test whether transparency about AI presence mitigates potential harm.
Results reveal that AI's corrupting force is as strong as humans'
arXiv Detail & Related papers (2021-02-15T13:15:12Z) - Human-Level Performance in No-Press Diplomacy via Equilibrium Search [29.858369754530905]
We describe an agent for the no-press variant of Diplomacy that combines supervised learning on human data with one-step lookahead search via regret minimization.
We show that our agent greatly exceeds the performance of past no-press Diplomacy bots, is unexploitable by expert humans, and ranks in the top 2% of human players when playing anonymous games on a popular Diplomacy website.
arXiv Detail & Related papers (2020-10-06T01:28:34Z) - Efficient exploration of zero-sum stochastic games [83.28949556413717]
We investigate the increasingly important and common game-solving setting where we do not have an explicit description of the game but only oracle access to it through gameplay.
During a limited-duration learning phase, the algorithm can control the actions of both players in order to try to learn the game and how to play it well.
Our motivation is to quickly learn strategies that have low exploitability in situations where evaluating the payoffs of a queried strategy profile is costly.
arXiv Detail & Related papers (2020-02-24T20:30:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.