Method for making multi-attribute decisions in wargames by combining
intuitionistic fuzzy numbers with reinforcement learning
- URL: http://arxiv.org/abs/2109.02354v1
- Date: Mon, 6 Sep 2021 10:45:52 GMT
- Title: Method for making multi-attribute decisions in wargames by combining
intuitionistic fuzzy numbers with reinforcement learning
- Authors: Yuxiang Sun, Bo Yuan, Yufan Xue, Jiawei Zhou, Xiaoyu Zhang and
Xianzhong Zhou
- Abstract summary: The article proposes an algorithm that combines the multi-attribute management and reinforcement learning methods.
It solves the problem of the agent's low rate of winning against specific rules and its inability to quickly converge during intelligent wargame training.
It is the first time in this field that an algorithm design for intelligent wargaming combines multi-attribute decision making with reinforcement learning.
- Score: 18.04026817707759
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Researchers are increasingly focusing on intelligent games as a hot research
area.The article proposes an algorithm that combines the multi-attribute
management and reinforcement learning methods, and that combined their effect
on wargaming, it solves the problem of the agent's low rate of winning against
specific rules and its inability to quickly converge during intelligent wargame
training.At the same time, this paper studied a multi-attribute decision making
and reinforcement learning algorithm in a wargame simulation environment, and
obtained data on red and blue conflict.Calculate the weight of each attribute
based on the intuitionistic fuzzy number weight calculations. Then determine
the threat posed by each opponent's chess pieces.Using the red side
reinforcement learning reward function, the AC framework is trained on the
reward function, and an algorithm combining multi-attribute decision-making
with reinforcement learning is obtained. A simulation experiment confirms that
the algorithm of multi-attribute decision-making combined with reinforcement
learning presented in this paper is significantly more intelligent than the
pure reinforcement learning algorithm.By resolving the shortcomings of the
agent's neural network, coupled with sparse rewards in large-map combat games,
this robust algorithm effectively reduces the difficulties of convergence. It
is also the first time in this field that an algorithm design for intelligent
wargaming combines multi-attribute decision making with reinforcement
learning.Attempt interdisciplinary cross-innovation in the academic field, like
designing intelligent wargames and improving reinforcement learning algorithms.
Related papers
- Reasoning, Memorization, and Fine-Tuning Language Models for Non-Cooperative Games [18.406992961818368]
We develop a method that integrates the tree of thoughts and multi-agent framework to enhance the capability of pre-trained language models in solving games.
We demonstrate a 65 percent winning rate against benchmark algorithms, with an additional 10 percent improvement after fine-tuning.
arXiv Detail & Related papers (2024-10-18T22:28:22Z) - Mastering Chinese Chess AI (Xiangqi) Without Search [2.309569018066392]
We have developed a high-performance Chinese Chess AI that operates without reliance on search algorithms.
This AI has demonstrated the capability to compete at a level commensurate with the top 0.1% of human players.
arXiv Detail & Related papers (2024-10-07T09:27:51Z) - A Unified Approach to Reinforcement Learning, Quantal Response
Equilibria, and Two-Player Zero-Sum Games [104.3339905200105]
This work studies an algorithm, which we call magnetic mirror descent, that is inspired by mirror descent and the non-Euclidean proximal gradient algorithm.
Our contribution is demonstrating the virtues of magnetic mirror descent as both an equilibrium solver and as an approach to reinforcement learning in two-player zero-sum games.
arXiv Detail & Related papers (2022-06-12T19:49:14Z) - No-Regret Learning in Time-Varying Zero-Sum Games [99.86860277006318]
Learning from repeated play in a fixed zero-sum game is a classic problem in game theory and online learning.
We develop a single parameter-free algorithm that simultaneously enjoys favorable guarantees under three performance measures.
Our algorithm is based on a two-layer structure with a meta-algorithm learning over a group of black-box base-learners satisfying a certain property.
arXiv Detail & Related papers (2022-01-30T06:10:04Z) - Online Learning in Budget-Constrained Dynamic Colonel Blotto Games [2.132096006921048]
We study the strategic allocation of limited resources using a Colonel Blotto game (CBG) under a dynamic setting.
We devise an efficient algorithm that combines a special bandit algorithm for path planning problem and a bandits with knapsack algorithm to cope with the budget constraint.
arXiv Detail & Related papers (2021-03-23T20:52:56Z) - Multi-Task Federated Reinforcement Learning with Adversaries [2.6080102941802106]
Reinforcement learning algorithms pose a serious threat from adversaries.
In this paper, we analyze the Multi-task Federated Reinforcement Learning algorithms.
We propose an adaptive attack method with better attack performance.
arXiv Detail & Related papers (2021-03-11T05:39:52Z) - Disturbing Reinforcement Learning Agents with Corrupted Rewards [62.997667081978825]
We analyze the effects of different attack strategies based on reward perturbations on reinforcement learning algorithms.
We show that smoothly crafting adversarial rewards are able to mislead the learner, and that using low exploration probability values, the policy learned is more robust to corrupt rewards.
arXiv Detail & Related papers (2021-02-12T15:53:48Z) - Evolving Reinforcement Learning Algorithms [186.62294652057062]
We propose a method for meta-learning reinforcement learning algorithms.
The learned algorithms are domain-agnostic and can generalize to new environments not seen during training.
We highlight two learned algorithms which obtain good generalization performance over other classical control tasks, gridworld type tasks, and Atari games.
arXiv Detail & Related papers (2021-01-08T18:55:07Z) - Learning to Play Sequential Games versus Unknown Opponents [93.8672371143881]
We consider a repeated sequential game between a learner, who plays first, and an opponent who responds to the chosen action.
We propose a novel algorithm for the learner when playing against an adversarial sequence of opponents.
Our results include algorithm's regret guarantees that depend on the regularity of the opponent's response.
arXiv Detail & Related papers (2020-07-10T09:33:05Z) - Provable Self-Play Algorithms for Competitive Reinforcement Learning [48.12602400021397]
We study self-play in competitive reinforcement learning under the setting of Markov games.
We show that a self-play algorithm achieves regret $tildemathcalO(sqrtT)$ after playing $T$ steps of the game.
We also introduce an explore-then-exploit style algorithm, which achieves a slightly worse regret $tildemathcalO(T2/3)$, but is guaranteed to run in time even in the worst case.
arXiv Detail & Related papers (2020-02-10T18:44:50Z) - Algorithms in Multi-Agent Systems: A Holistic Perspective from
Reinforcement Learning and Game Theory [2.5147566619221515]
Deep reinforcement learning has achieved outstanding results in recent years.
Recent works are exploring learning beyond single-agent scenarios and considering multi-agent scenarios.
Traditional game-theoretic algorithms, which, in turn, show bright application promise combined with modern algorithms and boosting computing power.
arXiv Detail & Related papers (2020-01-17T15:08:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.