Tricking AI chips into Simulating the Human Brain: A Detailed
Performance Analysis
- URL: http://arxiv.org/abs/2301.13637v1
- Date: Tue, 31 Jan 2023 13:51:37 GMT
- Title: Tricking AI chips into Simulating the Human Brain: A Detailed
Performance Analysis
- Authors: Lennart P. L. Landsmeer, Max C. W. Engelen, Rene Miedema and Christos
Strydis
- Abstract summary: We evaluate multiple, cutting-edge AI-chips (Graphcore IPU, GroqChip, Nvidia GPU with inferior Cores and Google TPU) for brain simulation.
Our performance analysis reveals that the simulation problem maps extremely well onto the GPU and TPU architectures.
The GroqChip outperforms both platforms for small networks but, due to implementing some floating-point operations at reduced accuracy, is found not yet usable for brain simulation.
- Score: 0.5354801701968198
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Challenging the Nvidia monopoly, dedicated AI-accelerator chips have begun
emerging for tackling the computational challenge that the inference and,
especially, the training of modern deep neural networks (DNNs) poses to modern
computers. The field has been ridden with studies assessing the performance of
these contestants across various DNN model types. However, AI-experts are aware
of the limitations of current DNNs and have been working towards the fourth AI
wave which will, arguably, rely on more biologically inspired models,
predominantly on spiking neural networks (SNNs). At the same time, GPUs have
been heavily used for simulating such models in the field of computational
neuroscience, yet AI-chips have not been tested on such workloads. The current
paper aims at filling this important gap by evaluating multiple, cutting-edge
AI-chips (Graphcore IPU, GroqChip, Nvidia GPU with Tensor Cores and Google TPU)
on simulating a highly biologically detailed model of a brain region, the
inferior olive (IO). This IO application stress-tests the different
AI-platforms for highlighting architectural tradeoffs by varying its compute
density, memory requirements and floating-point numerical accuracy. Our
performance analysis reveals that the simulation problem maps extremely well
onto the GPU and TPU architectures, which for networks of 125,000 cells leads
to a 28x respectively 1,208x speedup over CPU runtimes. At this speed, the TPU
sets a new record for largest real-time IO simulation. The GroqChip outperforms
both platforms for small networks but, due to implementing some floating-point
operations at reduced accuracy, is found not yet usable for brain simulation.
Related papers
- Training of Physical Neural Networks [16.832505972118646]
Physical neural networks (PNNs) are neural-like networks that leverage the properties of physical systems to perform computation.
PNNs could one day radically change what is possible and practical for AI systems.
To train PNNs at large scale, many methods including backpropagation-based and backpropagation-free approaches are now being explored.
arXiv Detail & Related papers (2024-06-05T15:28:04Z) - Fast Algorithms for Spiking Neural Network Simulation with FPGAs [0.0]
We create spiking neural network (SNN) simulators for the Potjans-Diesmann cortical microcircuit for a high-end Field-Programmable Gate Array (FPGA)
Our best simulators simulate the circuit 25% faster than real-time, require less than 21 nJ per synaptic event, and are bottle-necked by the device's on-chip memory.
This result is the first for simulating the circuit on a single hardware accelerator.
arXiv Detail & Related papers (2024-05-03T11:39:25Z) - Spyx: A Library for Just-In-Time Compiled Optimization of Spiking Neural
Networks [0.08965418284317034]
Spiking Neural Networks (SNNs) offer to enhance energy efficiency through a reduced and low-power hardware footprint.
This paper introduces Spyx, a new and lightweight SNN simulation and optimization library designed in JAX.
arXiv Detail & Related papers (2024-02-29T09:46:44Z) - Pruning random resistive memory for optimizing analogue AI [54.21621702814583]
AI models present unprecedented challenges to energy consumption and environmental sustainability.
One promising solution is to revisit analogue computing, a technique that predates digital computing.
Here, we report a universal solution, software-hardware co-design using structural plasticity-inspired edge pruning.
arXiv Detail & Related papers (2023-11-13T08:59:01Z) - Harnessing Manycore Processors with Distributed Memory for Accelerated
Training of Sparse and Recurrent Models [43.1773057439246]
Current AI training infrastructure is dominated by single instruction multiple data (SIMD) and systolic array architectures.
We explore sparse and recurrent model training on a massively parallel multiple instruction multiple data architecture with distributed local memory.
arXiv Detail & Related papers (2023-11-07T23:18:35Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Benchmarking GPU and TPU Performance with Graph Neural Networks [0.0]
This work analyzes and compares the GPU and TPU performance training a Graph Neural Network (GNN) developed to solve a real-life pattern recognition problem.
Characterizing the new class of models acting on sparse data may prove helpful in optimizing the design of deep learning libraries and future AI accelerators.
arXiv Detail & Related papers (2022-10-21T21:03:40Z) - Braille Letter Reading: A Benchmark for Spatio-Temporal Pattern
Recognition on Neuromorphic Hardware [50.380319968947035]
Recent deep learning approaches have reached accuracy in such tasks, but their implementation on conventional embedded solutions is still computationally very and energy expensive.
We propose a new benchmark for computing tactile pattern recognition at the edge through letters reading.
We trained and compared feed-forward and recurrent spiking neural networks (SNNs) offline using back-propagation through time with surrogate gradients, then we deployed them on the Intel Loihimorphic chip for efficient inference.
Our results show that the LSTM outperforms the recurrent SNN in terms of accuracy by 14%. However, the recurrent SNN on Loihi is 237 times more energy
arXiv Detail & Related papers (2022-05-30T14:30:45Z) - Dynamic Split Computing for Efficient Deep Edge Intelligence [78.4233915447056]
We introduce dynamic split computing, where the optimal split location is dynamically selected based on the state of the communication channel.
We show that dynamic split computing achieves faster inference in edge computing environments where the data rate and server load vary over time.
arXiv Detail & Related papers (2022-05-23T12:35:18Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - Mapping and Validating a Point Neuron Model on Intel's Neuromorphic
Hardware Loihi [77.34726150561087]
We investigate the potential of Intel's fifth generation neuromorphic chip - Loihi'
Loihi is based on the novel idea of Spiking Neural Networks (SNNs) emulating the neurons in the brain.
We find that Loihi replicates classical simulations very efficiently and scales notably well in terms of both time and energy performance as the networks get larger.
arXiv Detail & Related papers (2021-09-22T16:52:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.