From Clean Room to Machine Room: Commissioning of the First-Generation
BrainScaleS Wafer-Scale Neuromorphic System
- URL: http://arxiv.org/abs/2303.12359v1
- Date: Wed, 22 Mar 2023 07:50:51 GMT
- Title: From Clean Room to Machine Room: Commissioning of the First-Generation
BrainScaleS Wafer-Scale Neuromorphic System
- Authors: Hartmut Schmidt, Jos\'e Montes, Andreas Gr\"ubl, Maurice G\"uttler,
Dan Husmann, Joscha Ilmberger, Jakob Kaiser, Christian Mauch, Eric M\"uller,
Lars Sterzenbach, Johannes Schemmel, Sebastian Schmitt
- Abstract summary: BrainScaleS-1 is a neuromorphic system for emulating large-scale networks of spiking neurons.
A fault-tolerant design allows it to achieve wafer-scale integration despite unavoidable analog variability and component failures.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The first-generation of BrainScaleS, also referred to as BrainScaleS-1, is a
neuromorphic system for emulating large-scale networks of spiking neurons.
Following a "physical modeling" principle, its VLSI circuits are designed to
emulate the dynamics of biological examples: analog circuits implement neurons
and synapses with time constants that arise from their electronic components'
intrinsic properties. It operates in continuous time, with dynamics typically
matching an acceleration factor of 10000 compared to the biological regime. A
fault-tolerant design allows it to achieve wafer-scale integration despite
unavoidable analog variability and component failures. In this paper, we
present the commissioning process of a BrainScaleS-1 wafer module, providing a
short description of the system's physical components, illustrating the steps
taken during its assembly and the measures taken to operate it. Furthermore, we
reflect on the system's development process and the lessons learned to conclude
with a demonstration of its functionality by emulating a wafer-scale
synchronous firing chain, the largest spiking network emulation ran with analog
components and individual synapses to date.
Related papers
- Interpretable Spatio-Temporal Embedding for Brain Structural-Effective Network with Ordinary Differential Equation [56.34634121544929]
In this study, we first construct the brain-effective network via the dynamic causal model.
We then introduce an interpretable graph learning framework termed Spatio-Temporal Embedding ODE (STE-ODE)
This framework incorporates specifically designed directed node embedding layers, aiming at capturing the dynamic interplay between structural and effective networks.
arXiv Detail & Related papers (2024-05-21T20:37:07Z) - The Expressive Leaky Memory Neuron: an Efficient and Expressive Phenomenological Neuron Model Can Solve Long-Horizon Tasks [64.08042492426992]
We introduce the Expressive Memory (ELM) neuron model, a biologically inspired model of a cortical neuron.
Our ELM neuron can accurately match the aforementioned input-output relationship with under ten thousand trainable parameters.
We evaluate it on various tasks with demanding temporal structures, including the Long Range Arena (LRA) datasets.
arXiv Detail & Related papers (2023-06-14T13:34:13Z) - Contrastive-Signal-Dependent Plasticity: Forward-Forward Learning of
Spiking Neural Systems [73.18020682258606]
We develop a neuro-mimetic architecture, composed of spiking neuronal units, where individual layers of neurons operate in parallel.
We propose an event-based generalization of forward-forward learning, which we call contrastive-signal-dependent plasticity (CSDP)
Our experimental results on several pattern datasets demonstrate that the CSDP process works well for training a dynamic recurrent spiking network capable of both classification and reconstruction.
arXiv Detail & Related papers (2023-03-30T02:40:28Z) - Phenomenological Model of Superconducting Optoelectronic Loop Neurons [0.0]
Superconducting optoelectronic loop neurons are a class of circuits potentially conducive to networks for large-scale artificial cognition.
To date, all simulations of loop neurons have used first-principles circuit analysis to model the behavior of synapses, dendrites, and neurons.
Here we introduce a modeling framework that captures the behavior of the relevant synaptic, dendritic, and neuronal circuits.
arXiv Detail & Related papers (2022-10-18T16:38:35Z) - An accurate and flexible analog emulation of AdEx neuron dynamics in
silicon [0.0]
This manuscript presents the analog neuron circuits of the mixed-signal accelerated neuromorphic system BrainScaleS-2.
They are capable of flexibly and accurately emulating the adaptive exponential integrate-and-fire model equations in combination with both current- and conductance-based synapses.
arXiv Detail & Related papers (2022-09-19T18:08:23Z) - Mapping and Validating a Point Neuron Model on Intel's Neuromorphic
Hardware Loihi [77.34726150561087]
We investigate the potential of Intel's fifth generation neuromorphic chip - Loihi'
Loihi is based on the novel idea of Spiking Neural Networks (SNNs) emulating the neurons in the brain.
We find that Loihi replicates classical simulations very efficiently and scales notably well in terms of both time and energy performance as the networks get larger.
arXiv Detail & Related papers (2021-09-22T16:52:51Z) - Spatiotemporal Spike-Pattern Selectivity in Single Mixed-Signal Neurons
with Balanced Synapses [0.27998963147546135]
Mixed-signal neuromorphic processors could be used for inference and learning.
We show how inhomogeneous synaptic circuits could be utilized for resource-efficient implementation of network layers.
arXiv Detail & Related papers (2021-06-10T12:04:03Z) - Continuous Learning and Adaptation with Membrane Potential and
Activation Threshold Homeostasis [91.3755431537592]
This paper presents the Membrane Potential and Activation Threshold Homeostasis (MPATH) neuron model.
The model allows neurons to maintain a form of dynamic equilibrium by automatically regulating their activity when presented with input.
Experiments demonstrate the model's ability to adapt to and continually learn from its input.
arXiv Detail & Related papers (2021-04-22T04:01:32Z) - Emergence of Lie symmetries in functional architectures learned by CNNs [63.69764116066748]
We study the spontaneous development of symmetries in the early layers of a Convolutional Neural Network (CNN) during learning on natural images.
Our architecture is built in such a way to mimic the early stages of biological visual systems.
arXiv Detail & Related papers (2021-04-17T13:23:26Z) - Accelerated Analog Neuromorphic Computing [0.0]
This paper presents the concepts behind the BrainScales (BSS) accelerated analog neuromorphic computing architecture.
It describes the second-generation BrainScales-2 (BSS-2) version and its most recent in-silico realization, the HICANN-X Application Specific Integrated Circuit (ASIC)
The presented architecture is based upon a continuous-time, analog, physical model implementation of neurons and synapses.
arXiv Detail & Related papers (2020-03-26T16:00:55Z) - Structural plasticity on an accelerated analog neuromorphic hardware
system [0.46180371154032884]
We present a strategy to achieve structural plasticity by constantly rewiring the pre- and gpostsynaptic partners.
We implemented this algorithm on the analog neuromorphic system BrainScaleS-2.
We evaluated our implementation in a simple supervised learning scenario, showing its ability to optimize the network topology.
arXiv Detail & Related papers (2019-12-27T10:15:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.