Puzzle Solving without Search or Human Knowledge: An Unnatural Language
Approach
- URL: http://arxiv.org/abs/2109.02797v1
- Date: Tue, 7 Sep 2021 01:20:28 GMT
- Title: Puzzle Solving without Search or Human Knowledge: An Unnatural Language
Approach
- Authors: David Noever and Ryerson Burdick
- Abstract summary: The application of Generative Pre-trained Transformer (GPT-2) to learn text-archived game notation provides a model environment for exploring sparse reward gameplay.
The transformer architecture proves amenable to training on solved text archives describing mazes, Rubik's, and Sudoku solvers.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The application of Generative Pre-trained Transformer (GPT-2) to learn
text-archived game notation provides a model environment for exploring sparse
reward gameplay. The transformer architecture proves amenable to training on
solved text archives describing mazes, Rubik's Cube, and Sudoku solvers. The
method benefits from fine-tuning the transformer architecture to visualize
plausible strategies derived outside any guidance from human heuristics or
domain expertise. The large search space ($>10^{19}$) for the games provides a
puzzle environment in which the solution has few intermediate rewards and a
final move that solves the challenge.
Related papers
- Causal Language Modeling Can Elicit Search and Reasoning Capabilities on Logic Puzzles [14.777731830394357]
We study if causal language modeling can learn a complex task such as solving Sudoku puzzles.
We show that Transformer models trained on this synthetic task can indeed learn to solve Sudokus.
We also extend our analysis to Zebra puzzles (known as Einstein puzzles) and show that the model solves $92.04 %$ of the puzzles fully correctly.
arXiv Detail & Related papers (2024-09-16T17:42:15Z) - Solving Masked Jigsaw Puzzles with Diffusion Vision Transformers [5.374411622670979]
Image and video jigsaw puzzles pose the challenging task of rearranging image fragments or video frames from unordered sequences to restore meaningful images and video sequences.
Existing approaches often hinge on discriminative models tasked with predicting either the absolute positions of puzzle elements or the permutation actions applied to the original data.
We propose JPDVT, an innovative approach that harnesses diffusion transformers to address this challenge.
arXiv Detail & Related papers (2024-04-10T18:40:23Z) - Automated Graph Genetic Algorithm based Puzzle Validation for Faster
Game Desig [69.02688684221265]
This paper presents an evolutionary algorithm, empowered by expert-knowledge informeds, for solving logical puzzles in video games efficiently.
We discuss multiple variations of hybrid genetic approaches for constraint satisfaction problems that allow us to find a diverse set of near-optimal solutions for puzzles.
arXiv Detail & Related papers (2023-02-17T18:15:33Z) - Video Anomaly Detection by Solving Decoupled Spatio-Temporal Jigsaw
Puzzles [67.39567701983357]
Video Anomaly Detection (VAD) is an important topic in computer vision.
Motivated by the recent advances in self-supervised learning, this paper addresses VAD by solving an intuitive yet challenging pretext task.
Our method outperforms state-of-the-art counterparts on three public benchmarks.
arXiv Detail & Related papers (2022-07-20T19:49:32Z) - An AlphaZero-Inspired Approach to Solving Search Problems [63.24965775030674]
We adapt the methods and techniques used in AlphaZero for solving search problems.
We describe possible representations in terms of easy-instance solvers and self-reductions.
We also describe a version of Monte Carlo tree search adapted for search problems.
arXiv Detail & Related papers (2022-07-02T23:39:45Z) - CubeTR: Learning to Solve The Rubiks Cube Using Transformers [0.0]
The Rubiks cube has a single solved state for quintillions of possible configurations which leads to extremely sparse rewards.
The proposed model CubeTR attends to longer sequences of actions and addresses the problem of sparse rewards.
arXiv Detail & Related papers (2021-11-11T03:17:28Z) - Using Small MUSes to Explain How to Solve Pen and Paper Puzzles [4.535832029902474]
We present Demystify, a tool which allows puzzles to be expressed in a high-level constraint programming language.
We give several improvements to the existing techniques for solving puzzles with MUSes.
We demonstrate the effectiveness and generality of Demystify by comparing its results to documented strategies for solving a range of pen and paper puzzles by hand.
arXiv Detail & Related papers (2021-04-30T15:07:51Z) - Non-Rigid Puzzles [50.213265511586535]
We present a non-rigid multi-part shape matching algorithm.
We assume to be given a reference shape and its multiple parts undergoing a non-rigid deformation.
Experimental results on synthetic as well as real scans demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2020-11-26T00:32:30Z) - Interactive Fiction Game Playing as Multi-Paragraph Reading
Comprehension with Reinforcement Learning [94.50608198582636]
Interactive Fiction (IF) games with real human-written natural language texts provide a new natural evaluation for language understanding techniques.
We take a novel perspective of IF game solving and re-formulate it as Multi-Passage Reading (MPRC) tasks.
arXiv Detail & Related papers (2020-10-05T23:09:20Z) - Exploration Based Language Learning for Text-Based Games [72.30525050367216]
This work presents an exploration and imitation-learning-based agent capable of state-of-the-art performance in playing text-based computer games.
Text-based computer games describe their world to the player through natural language and expect the player to interact with the game using text.
These games are of interest as they can be seen as a testbed for language understanding, problem-solving, and language generation by artificial agents.
arXiv Detail & Related papers (2020-01-24T03:03:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.