TxSim:Modeling Training of Deep Neural Networks on Resistive Crossbar
Systems
- URL: http://arxiv.org/abs/2002.11151v3
- Date: Fri, 8 Jan 2021 03:54:43 GMT
- Title: TxSim:Modeling Training of Deep Neural Networks on Resistive Crossbar
Systems
- Authors: Sourjya Roy, Shrihari Sridharan, Shubham Jain, and Anand Raghunathan
- Abstract summary: crossbar-based computations face a major challenge due to a variety of device and circuit-level non-idealities.
We propose TxSim, a fast and customizable modeling framework to functionally evaluate DNN training on crossbar-based hardware.
- Score: 3.1887081453726136
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Resistive crossbars have attracted significant interest in the design of Deep
Neural Network (DNN) accelerators due to their ability to natively execute
massively parallel vector-matrix multiplications within dense memory arrays.
However, crossbar-based computations face a major challenge due to a variety of
device and circuit-level non-idealities, which manifest as errors in the
vector-matrix multiplications and eventually degrade DNN accuracy. To address
this challenge, there is a need for tools that can model the functional impact
of non-idealities on DNN training and inference. Existing efforts towards this
goal are either limited to inference, or are too slow to be used for
large-scale DNN training. We propose TxSim, a fast and customizable modeling
framework to functionally evaluate DNN training on crossbar-based hardware
considering the impact of non-idealities. The key features of TxSim that
differentiate it from prior efforts are: (i) It comprehensively models
non-idealities during all training operations (forward propagation, backward
propagation, and weight update) and (ii) it achieves computational efficiency
by mapping crossbar evaluations to well-optimized BLAS routines and
incorporates speedup techniques to further reduce simulation time with minimal
impact on accuracy. TxSim achieves orders-of-magnitude improvement in
simulation speed over prior works, and thereby makes it feasible to evaluate
training of large-scale DNNs on crossbars. Our experiments using TxSim reveal
that the accuracy degradation in DNN training due to non-idealities can be
substantial (3%-10%) for large-scale DNNs, underscoring the need for further
research in mitigation techniques. We also analyze the impact of various device
and circuit-level parameters and the associated non-idealities to provide key
insights that can guide the design of crossbar-based DNN training accelerators.
Related papers
- Hardware-aware training for large-scale and diverse deep learning
inference workloads using in-memory computing-based accelerators [7.152059921639833]
We show that many large-scale deep neural networks can be successfully retrained to show iso-accuracy on AIMC.
Our results suggest that AIMC nonidealities that add noise to the inputs or outputs, not the weights, have the largest impact on DNN accuracy.
arXiv Detail & Related papers (2023-02-16T18:25:06Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Online Training Through Time for Spiking Neural Networks [66.7744060103562]
Spiking neural networks (SNNs) are promising brain-inspired energy-efficient models.
Recent progress in training methods has enabled successful deep SNNs on large-scale tasks with low latency.
We propose online training through time (OTTT) for SNNs, which is derived from BPTT to enable forward-in-time learning.
arXiv Detail & Related papers (2022-10-09T07:47:56Z) - Recurrent Bilinear Optimization for Binary Neural Networks [58.972212365275595]
BNNs neglect the intrinsic bilinear relationship of real-valued weights and scale factors.
Our work is the first attempt to optimize BNNs from the bilinear perspective.
We obtain robust RBONNs, which show impressive performance over state-of-the-art BNNs on various models and datasets.
arXiv Detail & Related papers (2022-09-04T06:45:33Z) - Training High-Performance Low-Latency Spiking Neural Networks by
Differentiation on Spike Representation [70.75043144299168]
Spiking Neural Network (SNN) is a promising energy-efficient AI model when implemented on neuromorphic hardware.
It is a challenge to efficiently train SNNs due to their non-differentiability.
We propose the Differentiation on Spike Representation (DSR) method, which could achieve high performance.
arXiv Detail & Related papers (2022-05-01T12:44:49Z) - Shift-BNN: Highly-Efficient Probabilistic Bayesian Neural Network
Training via Memory-Friendly Pattern Retrieving [5.043640793217879]
We design and prototype the first highly efficient BNN training accelerator, named Shift-BNN, that is low-cost and scalable.
Shift-BNN achieves an average of 4.9x (up to 10.8x) boost in energy efficiency and 1.6x (up to 2.8x) speedup over the baseline DNN training accelerator.
arXiv Detail & Related papers (2021-10-07T15:20:53Z) - slimTrain -- A Stochastic Approximation Method for Training Separable
Deep Neural Networks [2.4373900721120285]
DeepTrain networks (DNNs) have shown their success as high-dimensional neural function approximators in many applications.
We propose slimTrain, a modest optimization method for training DNNs with reduced sensitivity to the choice hyper-dimensional datasets.
arXiv Detail & Related papers (2021-09-28T19:31:57Z) - Low-Precision Training in Logarithmic Number System using Multiplicative
Weight Update [49.948082497688404]
Training large-scale deep neural networks (DNNs) currently requires a significant amount of energy, leading to serious environmental impacts.
One promising approach to reduce the energy costs is representing DNNs with low-precision numbers.
We jointly design a lowprecision training framework involving a logarithmic number system (LNS) and a multiplicative weight update training method, termed LNS-Madam.
arXiv Detail & Related papers (2021-06-26T00:32:17Z) - FracTrain: Fractionally Squeezing Bit Savings Both Temporally and
Spatially for Efficient DNN Training [81.85361544720885]
We propose FracTrain that integrates progressive fractional quantization which gradually increases the precision of activations, weights, and gradients.
FracTrain reduces computational cost and hardware-quantified energy/latency of DNN training while achieving a comparable or better (-0.12%+1.87%) accuracy.
arXiv Detail & Related papers (2020-12-24T05:24:10Z) - Procrustes: a Dataflow and Accelerator for Sparse Deep Neural Network
Training [0.5219568203653523]
We develop a sparse DNN training accelerator that produces pruned models with the same accuracy as dense models without first training, then pruning, and finally retraining, a dense model.
Compared to training the equivalent unpruned models using a state-of-the-art DNN accelerator without sparse training support, Procrustes consumes up to 3.26$times$ less energy and offers up to 4$times$ speedup across a range of models, while pruning weights by an order of magnitude and maintaining unpruned accuracy.
arXiv Detail & Related papers (2020-09-23T07:39:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.