Distributed Compressed Sparse Row Format for Spiking Neural Network
Simulation, Serialization, and Interoperability
- URL: http://arxiv.org/abs/2304.05587v1
- Date: Wed, 12 Apr 2023 03:19:06 GMT
- Title: Distributed Compressed Sparse Row Format for Spiking Neural Network
Simulation, Serialization, and Interoperability
- Authors: Felix Wang
- Abstract summary: We discuss a parallel extension of a widely used format for efficiently representing sparse matrices, the compressed sparse row (CSR)
We contend that organizing additional network information, such as neuron and synapse state, in alignment with its adjacency as dCSR provides a straightforward partition-based distribution of network state.
We provide a potential implementation, and put it forward for adoption within the neural computing community.
- Score: 0.48733623015338234
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the increasing development of neuromorphic platforms and their related
software tools as well as the increasing scale of spiking neural network (SNN)
models, there is a pressure for interoperable and scalable representations of
network state. In response to this, we discuss a parallel extension of a widely
used format for efficiently representing sparse matrices, the compressed sparse
row (CSR), in the context of supporting the simulation and serialization of
large-scale SNNs. Sparse matrices for graph adjacency structure provide a
natural fit for describing the connectivity of an SNN, and prior work in the
area of parallel graph partitioning has developed the distributed CSR (dCSR)
format for storing and ingesting large graphs. We contend that organizing
additional network information, such as neuron and synapse state, in alignment
with its adjacency as dCSR provides a straightforward partition-based
distribution of network state. For large-scale simulations, this means each
parallel process is only responsible for its own partition of state, which
becomes especially useful when the size of an SNN exceeds the memory resources
of a single compute node. For potentially long-running simulations, this also
enables network serialization to and from disk (e.g. for checkpoint/restart
fault-tolerant computing) to be performed largely independently between
parallel processes. We also provide a potential implementation, and put it
forward for adoption within the neural computing community.
Related papers
- Heterogenous Memory Augmented Neural Networks [84.29338268789684]
We introduce a novel heterogeneous memory augmentation approach for neural networks.
By introducing learnable memory tokens with attention mechanism, we can effectively boost performance without huge computational overhead.
We show our approach on various image and graph-based tasks under both in-distribution (ID) and out-of-distribution (OOD) conditions.
arXiv Detail & Related papers (2023-10-17T01:05:28Z) - Fast inference of latent space dynamics in huge relational event
networks [0.0]
We propose a likelihood-based algorithm that can deal with huge event networks.
In this work we propose a hierarchical strategy for inferring network community dynamics embedded into an interpretable latent space.
To make the framework feasible for large networks we borrow from machine learning optimization methodology.
arXiv Detail & Related papers (2023-03-29T15:18:56Z) - Symbolic Synthesis of Neural Networks [0.0]
I present Graph-basedally Synthesized Neural Networks (GSSNNs)
GSSNNs are a form of neural network whose topology and parameters are informed by the output of a symbolic program.
I demonstrate that by developing symbolic abstractions at a population level, I can elicit reliable patterns of improved generalization with small quantities of data known to contain local and discrete features.
arXiv Detail & Related papers (2023-03-06T18:13:14Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Receptive Field-based Segmentation for Distributed CNN Inference
Acceleration in Collaborative Edge Computing [93.67044879636093]
We study inference acceleration using distributed convolutional neural networks (CNNs) in collaborative edge computing network.
We propose a novel collaborative edge computing using fused-layer parallelization to partition a CNN model into multiple blocks of convolutional layers.
arXiv Detail & Related papers (2022-07-22T18:38:11Z) - DistIR: An Intermediate Representation and Simulator for Efficient
Neural Network Distribution [15.086401550425125]
DistIR is a representation for distributed computation that is tailored for efficient analyses.
We show how DistIR and its simulator enable fast grid searches over complex distribution spaces spanning up to 1000+ configurations.
arXiv Detail & Related papers (2021-11-09T21:32:51Z) - Parallel Simulation of Quantum Networks with Distributed Quantum State
Management [56.24769206561207]
We identify requirements for parallel simulation of quantum networks and develop the first parallel discrete event quantum network simulator.
Our contributions include the design and development of a quantum state manager that maintains shared quantum information distributed across multiple processes.
We release the parallel SeQUeNCe simulator as an open-source tool alongside the existing sequential version.
arXiv Detail & Related papers (2021-11-06T16:51:17Z) - Partitioning sparse deep neural networks for scalable training and
inference [8.282177703075453]
State-of-the-art deep neural networks (DNNs) have significant computational and data management requirements.
Sparsification and pruning methods are shown to be effective in removing a large fraction of connections in DNNs.
The resulting sparse networks present unique challenges to further improve the computational efficiency of training and inference in deep learning.
arXiv Detail & Related papers (2021-04-23T20:05:52Z) - Combining Differentiable PDE Solvers and Graph Neural Networks for Fluid
Flow Prediction [79.81193813215872]
We develop a hybrid (graph) neural network that combines a traditional graph convolutional network with an embedded differentiable fluid dynamics simulator inside the network itself.
We show that we can both generalize well to new situations and benefit from the substantial speedup of neural network CFD predictions.
arXiv Detail & Related papers (2020-07-08T21:23:19Z) - A Linear Algebraic Approach to Model Parallelism in Deep Learning [0.0]
Training deep neural networks (DNNs) in large-cluster computing environments is increasingly necessary, as networks grow in size and complexity.
We propose a linear-algebraic approach to model parallelism in deep learning, which allows parallel distribution of any tensor in the DNN.
We build distributed DNN layers using these parallel primitives, composed with sequential layer implementations, and demonstrate their application by building and training a distributed DNN using DistDL, a PyTorch and MPI-based distributed deep learning toolkit.
arXiv Detail & Related papers (2020-06-04T19:38:05Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.