A Simulation Environment for the Neuroevolution of Ant Colony Dynamics
- URL: http://arxiv.org/abs/2406.13147v3
- Date: Mon, 21 Oct 2024 02:15:42 GMT
- Title: A Simulation Environment for the Neuroevolution of Ant Colony Dynamics
- Authors: Michael Crosscombe, Ilya Horiguchi, Norihiro Maruyama, Shigeto Dobata, Takashi Ikegami,
- Abstract summary: We introduce a simulation environment to facilitate research into emergent collective behaviour.
By leveraging real-world data, the environment simulates a target ant trail that a controllable agent must learn to replicate.
- Score: 0.0
- License:
- Abstract: We introduce a simulation environment to facilitate research into emergent collective behaviour, with a focus on replicating the dynamics of ant colonies. By leveraging real-world data, the environment simulates a target ant trail that a controllable agent must learn to replicate, using sensory data observed by the target ant. This work aims to contribute to the neuroevolution of models for collective behaviour, focusing on evolving neural architectures that encode domain-specific behaviours in the network topology. By evolving models that can be modified and studied in a controlled environment, we can uncover the necessary conditions required for collective behaviours to emerge. We hope this environment will be useful to those studying the role of interactions in emergent behaviour within collective systems.
Related papers
- Evolving Neural Networks Reveal Emergent Collective Behavior from Minimal Agent Interactions [0.0]
We investigate how neural networks evolve to control agents' behavior in a dynamic environment.
Simpler behaviors, such as lane formation and laminar flow, are characterized by more linear network operations.
Specific environmental parameters, such as moderate noise, broader field of view, and lower agent density, promote the evolution of non-linear networks.
arXiv Detail & Related papers (2024-10-25T17:43:00Z) - Navigating the swarm: Deep neural networks command emergent behaviours [2.7059353835118602]
We show that it is possible to generate coordinated structures in collective behavior with intended global patterns by fine-tuning an inter-agent interaction rule.
Our strategy employs deep neural networks, obeying the laws of dynamics, to find interaction rules that command desired structures.
Our findings pave the way for new applications in robotic swarm operations, active matter organisation, and for the uncovering of obscure interaction rules in biological systems.
arXiv Detail & Related papers (2024-07-16T02:46:11Z) - Cognitive Evolutionary Learning to Select Feature Interactions for Recommender Systems [59.117526206317116]
We show that CELL can adaptively evolve into different models for different tasks and data.
Experiments on four real-world datasets demonstrate that CELL significantly outperforms state-of-the-art baselines.
arXiv Detail & Related papers (2024-05-29T02:35:23Z) - Contrastive-Signal-Dependent Plasticity: Self-Supervised Learning in Spiking Neural Circuits [61.94533459151743]
This work addresses the challenge of designing neurobiologically-motivated schemes for adjusting the synapses of spiking networks.
Our experimental simulations demonstrate a consistent advantage over other biologically-plausible approaches when training recurrent spiking networks.
arXiv Detail & Related papers (2023-03-30T02:40:28Z) - Co-Imitation: Learning Design and Behaviour by Imitation [10.40773958250192]
Co-adaptation of robots aims to adapt both body and behaviour of a system for a given task.
This paper introduces a new viewpoint on the co-adaptation problem, which we call co-imitation.
We propose a co-imitation methodology for adapting behaviour and morphology by matching state distributions of the demonstrator.
arXiv Detail & Related papers (2022-09-02T17:57:32Z) - Towards the Neuroevolution of Low-level Artificial General Intelligence [5.2611228017034435]
We argue that the search for Artificial General Intelligence (AGI) should start from a much lower level than human-level intelligence.
Our hypothesis is that learning occurs through sensory feedback when an agent acts in an environment.
We evaluate a method to evolve a biologically-inspired artificial neural network that learns from environment reactions.
arXiv Detail & Related papers (2022-07-27T15:30:50Z) - Information is Power: Intrinsic Control via Information Capture [110.3143711650806]
We argue that a compact and general learning objective is to minimize the entropy of the agent's state visitation estimated using a latent state-space model.
This objective induces an agent to both gather information about its environment, corresponding to reducing uncertainty, and to gain control over its environment, corresponding to reducing the unpredictability of future world states.
arXiv Detail & Related papers (2021-12-07T18:50:42Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z) - Epigenetic evolution of deep convolutional models [81.21462458089142]
We build upon a previously proposed neuroevolution framework to evolve deep convolutional models.
We propose a convolutional layer layout which allows kernels of different shapes and sizes to coexist within the same layer.
The proposed layout enables the size and shape of individual kernels within a convolutional layer to be evolved with a corresponding new mutation operator.
arXiv Detail & Related papers (2021-04-12T12:45:16Z) - Ecological Reinforcement Learning [76.9893572776141]
We study the kinds of environment properties that can make learning under such conditions easier.
understanding how properties of the environment impact the performance of reinforcement learning agents can help us to structure our tasks in ways that make learning tractable.
arXiv Detail & Related papers (2020-06-22T17:55:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.