Playing a 2D Game Indefinitely using NEAT and Reinforcement Learning
- URL: http://arxiv.org/abs/2207.14140v1
- Date: Thu, 28 Jul 2022 15:01:26 GMT
- Title: Playing a 2D Game Indefinitely using NEAT and Reinforcement Learning
- Authors: Jerin Paul Selvan, Pravin S. Game
- Abstract summary: The performance of algorithms can be compared by using artificial agents that will behave according to the algorithm in the environment they are put in.
The algorithms that are enforced on the artificial agents are NeuroEvolution of Augmenting Topologies (NEAT) and Reinforcement Learning.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: For over a decade now, robotics and the use of artificial agents have become
a common thing.Testing the performance of new path finding or search space
optimization algorithms has also become a challenge as they require simulation
or an environment to test them.The creation of artificial environments with
artificial agents is one of the methods employed to test such algorithms.Games
have also become an environment to test them.The performance of the algorithms
can be compared by using artificial agents that will behave according to the
algorithm in the environment they are put in.The performance parameters can be,
how quickly the agent is able to differentiate between rewarding actions and
hostile actions.This can be tested by placing the agent in an environment with
different types of hurdles and the goal of the agent is to reach the farthest
by taking decisions on actions that will lead to avoiding all the obstacles.The
environment chosen is a game called "Flappy Bird".The goal of the game is to
make the bird fly through a set of pipes of random heights.The bird must go in
between these pipes and must not hit the top, the bottom, or the pipes
themselves.The actions that the bird can take are either to flap its wings or
drop down with gravity.The algorithms that are enforced on the artificial
agents are NeuroEvolution of Augmenting Topologies (NEAT) and Reinforcement
Learning.The NEAT algorithm takes an "N" initial population of artificial
agents.They follow genetic algorithms by considering an objective function,
crossover, mutation, and augmenting topologies.Reinforcement learning, on the
other hand, remembers the state, the action taken at that state, and the reward
received for the action taken using a single agent and a Deep Q-learning
Network.The performance of the NEAT algorithm improves as the initial
population of the artificial agents is increased.
Related papers
- AI planning in the imagination: High-level planning on learned abstract
search spaces [68.75684174531962]
We propose a new method, called PiZero, that gives an agent the ability to plan in an abstract search space that the agent learns during training.
We evaluate our method on multiple domains, including the traveling salesman problem, Sokoban, 2048, the facility location problem, and Pacman.
arXiv Detail & Related papers (2023-08-16T22:47:16Z) - Achieving mouse-level strategic evasion performance using real-time
computational planning [59.60094442546867]
Planning is an extraordinary ability in which the brain imagines and then enacts evaluated possible futures.
We develop a more efficient biologically-inspired planning algorithm, TLPPO, based on work on how the ecology of an animal governs the value of spatial planning.
We compare the performance of a real-time agent using TLPPO against the performance of live mice, all tasked with evading a robot predator.
arXiv Detail & Related papers (2022-11-04T18:34:36Z) - A neural net architecture based on principles of neural plasticity and
development evolves to effectively catch prey in a simulated environment [2.834895018689047]
A profound challenge for A-Life is to construct agents whose behavior is 'life-like' in a deep way.
We propose an architecture and approach to constructing networks driving artificial agents, using processes analogous to the processes that construct and sculpt the brains of animals.
We think this architecture may be useful for controlling small autonomous robots or drones, because it allows for a rapid response to changes in sensor inputs.
arXiv Detail & Related papers (2022-01-28T05:10:56Z) - Mimicking Playstyle by Adapting Parameterized Behavior Trees in RTS
Games [0.0]
Behavior Trees (BTs) impacted the field of Artificial Intelligence (AI) in games.
BTs forced complexity of handcrafted BTs to became barely-tractable and error-prone.
Recent trends in the field focused on automatic creation of AI-agents.
We present a novel approach to semi-automatic construction of AI-agents, that mimic and generalize given human gameplays.
arXiv Detail & Related papers (2021-11-23T20:36:28Z) - XAI-N: Sensor-based Robot Navigation using Expert Policies and Decision
Trees [55.9643422180256]
We present a novel sensor-based learning navigation algorithm to compute a collision-free trajectory for a robot in dense and dynamic environments.
Our approach uses deep reinforcement learning-based expert policy that is trained using a sim2real paradigm.
We highlight the benefits of our algorithm in simulated environments and navigating a Clearpath Jackal robot among moving pedestrians.
arXiv Detail & Related papers (2021-04-22T01:33:10Z) - Learning What To Do by Simulating the Past [76.86449554580291]
We show that by combining a learned feature encoder with learned inverse models, we can enable agents to simulate human actions backwards in time to infer what they must have done.
The resulting algorithm is able to reproduce a specific skill in MuJoCo environments given a single state sampled from the optimal policy for that skill.
arXiv Detail & Related papers (2021-04-08T17:43:29Z) - PlasticineLab: A Soft-Body Manipulation Benchmark with Differentiable
Physics [89.81550748680245]
We introduce a new differentiable physics benchmark called PasticineLab.
In each task, the agent uses manipulators to deform the plasticine into the desired configuration.
We evaluate several existing reinforcement learning (RL) methods and gradient-based methods on this benchmark.
arXiv Detail & Related papers (2021-04-07T17:59:23Z) - Learning Human Rewards by Inferring Their Latent Intelligence Levels in
Multi-Agent Games: A Theory-of-Mind Approach with Application to Driving Data [18.750834997334664]
We argue that humans are bounded rational and have different intelligence levels when reasoning about others' decision-making process.
We propose a new multi-agent Inverse Reinforcement Learning framework that reasons about humans' latent intelligence levels during learning.
arXiv Detail & Related papers (2021-03-07T07:48:31Z) - Accelerated Sim-to-Real Deep Reinforcement Learning: Learning Collision
Avoidance from Human Player [5.960346570280513]
This paper presents a sensor-level mapless collision avoidance algorithm for use in mobile robots.
An efficient training strategy is proposed to allow a robot to learn from both human experience data and self-exploratory data.
A game format simulation framework is designed to allow the human player to tele-operate the mobile robot to a goal.
arXiv Detail & Related papers (2021-02-21T23:27:34Z) - Generating Human-Like Movement: A Comparison Between Two Approaches
Based on Environmental Features [4.511923587827301]
Two novel algorithms have been presented to generate human-like trajectories based on environmental features.
The human-likeness aspect has been tested by a human expert judging the final generated trajectories as realistic.
We show how, despite generating trajectories closer to the real one according to our predefined metrics, the Feature-Based A* algorithm fall short in time efficiency compared to the Attraction-Based A* algorithm.
arXiv Detail & Related papers (2020-12-11T16:45:32Z) - Never Give Up: Learning Directed Exploration Strategies [63.19616370038824]
We propose a reinforcement learning agent to solve hard exploration games by learning a range of directed exploratory policies.
We construct an episodic memory-based intrinsic reward using k-nearest neighbors over the agent's recent experience to train the directed exploratory policies.
A self-supervised inverse dynamics model is used to train the embeddings of the nearest neighbour lookup, biasing the novelty signal towards what the agent can control.
arXiv Detail & Related papers (2020-02-14T13:57:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.