Finding the optimal human strategy for Wordle using maximum correct
letter probabilities and reinforcement learning
- URL: http://arxiv.org/abs/2202.00557v1
- Date: Tue, 1 Feb 2022 17:03:26 GMT
- Title: Finding the optimal human strategy for Wordle using maximum correct
letter probabilities and reinforcement learning
- Authors: Benton J. Anderson, Jesse G. Meyer
- Abstract summary: Wordle is an online word puzzle game that gained viral popularity in January 2022.
We present two different methods for choosing starting words along with a framework for discovering the optimal human strategy.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Wordle is an online word puzzle game that gained viral popularity in January
2022. The goal is to guess a hidden five letter word. After each guess, the
player gains information about whether the letters they guessed are present in
the word, and whether they are in the correct position. Numerous blogs have
suggested guessing strategies and starting word lists that improve the chance
of winning. Optimized algorithms can win 100% of games within five of the six
allowed trials. However, it is infeasible for human players to use these
algorithms due to an inability to perfectly recall all known 5-letter words and
perform complex calculations that optimize information gain. Here, we present
two different methods for choosing starting words along with a framework for
discovering the optimal human strategy based on reinforcement learning. Human
Wordle players can use the rules we discover to optimize their chance of
winning.
Related papers
- Wordle: A Microcosm of Life. Luck, Skill, Cheating, Loyalty, and
Influence! [0.0]
Wordle is a popular, online word game offered by the New York Times.
Players have 6 attempts to guess the daily word (target word)
After each attempt, the player receives color-coded information about the correctness and position of each letter in the guess.
arXiv Detail & Related papers (2023-09-05T10:38:53Z) - On Optimal Strategies for Wordle and General Guessing Games [6.85316573653194]
We develop a method for finding optimal strategies for guessing games while avoiding an exhaustive search.
This work is developed to apply to any guessing game, but we use Wordle as an example to present concrete results.
arXiv Detail & Related papers (2023-05-16T02:30:10Z) - ApproxED: Approximate exploitability descent via learned best responses [61.17702187957206]
We study the problem of finding an approximate Nash equilibrium of games with continuous action sets.
We propose two new methods that minimize an approximation of exploitability with respect to the strategy profile.
arXiv Detail & Related papers (2023-01-20T23:55:30Z) - Strategizing against Learners in Bayesian Games [74.46970859427907]
We study repeated two-player games where one of the players, the learner, employs a no-regret learning strategy.
We consider general Bayesian games, where the payoffs of both the payoffs of both the learner and the learner could depend on the type.
arXiv Detail & Related papers (2022-05-17T18:10:25Z) - Using Wordle for Learning to Design and Compare Strategies [0.685316573653194]
We can design parameterized strategies for solving Wordle, based on probabilistic, statistical, and information-theoretical information about the games.
The strategies can handle a reasonably large family of Wordle-like games both systematically and dynamically.
This paper will provide the results of using two families of parameterized strategies to solve the current Wordle.
arXiv Detail & Related papers (2022-04-30T14:41:25Z) - Selecting Seed Words for Wordle using Character Statistics [0.3108011671896571]
Wordle, a word guessing game rose to global popularity in the January of 2022.
The goal of the game is to guess a five-letter English word within six tries.
This study uses character statistics of five-letter words to determine the best three starting words.
arXiv Detail & Related papers (2022-02-07T19:01:19Z) - Semantic-Preserving Adversarial Text Attacks [85.32186121859321]
We propose a Bigram and Unigram based adaptive Semantic Preservation Optimization (BU-SPO) method to examine the vulnerability of deep models.
Our method achieves the highest attack success rates and semantics rates by changing the smallest number of words compared with existing methods.
arXiv Detail & Related papers (2021-08-23T09:05:18Z) - Playing Codenames with Language Graphs and Word Embeddings [21.358501003335977]
We propose an algorithm that can generate Codenames clues from the language graph BabelNet.
We introduce a new scoring function that measures the quality of clues.
We develop BabelNet-Word Selection Framework (BabelNet-WSF) to improve BabelNet clue quality.
arXiv Detail & Related papers (2021-05-12T18:23:03Z) - Faster Algorithms for Optimal Ex-Ante Coordinated Collusive Strategies
in Extensive-Form Zero-Sum Games [123.76716667704625]
We focus on the problem of finding an optimal strategy for a team of two players that faces an opponent in an imperfect-information zero-sum extensive-form game.
In that setting, it is known that the best the team can do is sample a profile of potentially randomized strategies (one per player) from a joint (a.k.a. correlated) probability distribution at the beginning of the game.
We provide an algorithm that computes such an optimal distribution by only using profiles where only one of the team members gets to randomize in each profile.
arXiv Detail & Related papers (2020-09-21T17:51:57Z) - Learning to Play Sequential Games versus Unknown Opponents [93.8672371143881]
We consider a repeated sequential game between a learner, who plays first, and an opponent who responds to the chosen action.
We propose a novel algorithm for the learner when playing against an adversarial sequence of opponents.
Our results include algorithm's regret guarantees that depend on the regularity of the opponent's response.
arXiv Detail & Related papers (2020-07-10T09:33:05Z) - Optimally Deceiving a Learning Leader in Stackelberg Games [123.14187606686006]
Recent results in the ML community have revealed that learning algorithms used to compute the optimal strategy for the leader to commit to in a Stackelberg game, are susceptible to manipulation by the follower.
This paper shows that it is always possible for the follower to compute (near-optimal) payoffs for various scenarios about the learning interaction between leader and follower.
arXiv Detail & Related papers (2020-06-11T16:18:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.