Binary stochasticity enabled highly efficient neuromorphic deep learning
achieves better-than-software accuracy
- URL: http://arxiv.org/abs/2304.12866v1
- Date: Tue, 25 Apr 2023 14:38:36 GMT
- Title: Binary stochasticity enabled highly efficient neuromorphic deep learning
achieves better-than-software accuracy
- Authors: Yang Li, Wei Wang, Ming Wang, Chunmeng Dou, Zhengyu Ma, Huihui Zhou,
Peng Zhang, Nicola Lepri, Xumeng Zhang, Qing Luo, Xiaoxin Xu, Guanhua Yang,
Feng Zhang, Ling Li, Daniele Ielmini, and Ming Liu
- Abstract summary: Deep learning needs high-precision handling of forwarding signals, backpropagating errors, and updating weights.
It is challenging to implement deep learning in hardware systems that use noisy analog memristors as artificial synapses.
We propose a binary learning algorithm that modifies all elementary neural network operations.
- Score: 17.11946381948498
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Deep learning needs high-precision handling of forwarding signals,
backpropagating errors, and updating weights. This is inherently required by
the learning algorithm since the gradient descent learning rule relies on the
chain product of partial derivatives. However, it is challenging to implement
deep learning in hardware systems that use noisy analog memristors as
artificial synapses, as well as not being biologically plausible.
Memristor-based implementations generally result in an excessive cost of
neuronal circuits and stringent demands for idealized synaptic devices. Here,
we demonstrate that the requirement for high precision is not necessary and
that more efficient deep learning can be achieved when this requirement is
lifted. We propose a binary stochastic learning algorithm that modifies all
elementary neural network operations, by introducing (i) stochastic
binarization of both the forwarding signals and the activation function
derivatives, (ii) signed binarization of the backpropagating errors, and (iii)
step-wised weight updates. Through an extensive hybrid approach of software
simulation and hardware experiments, we find that binary stochastic deep
learning systems can provide better performance than the software-based
benchmarks using the high-precision learning algorithm. Also, the binary
stochastic algorithm strongly simplifies the neural network operations in
hardware, resulting in an improvement of the energy efficiency for the
multiply-and-accumulate operations by more than three orders of magnitudes.
Related papers
- Semi-adaptive Synergetic Two-way Pseudoinverse Learning System [8.16000189123978]
We propose a semi-adaptive synergetic two-way pseudoinverse learning system.
Each subsystem encompasses forward learning, backward learning, and feature concatenation modules.
The whole system is trained using a non-gradient descent learning algorithm.
arXiv Detail & Related papers (2024-06-27T06:56:46Z) - BOLD: Boolean Logic Deep Learning [1.4272256806865107]
We introduce the notion of Boolean variation such that neurons made of Boolean weights and inputs can be trained efficiently in Boolean domain using Boolean logic instead of descent gradient and real arithmetic.
Our approach achieves baseline full-precision accuracy in ImageNet classification and surpasses state-of-the-art results in semantic segmentation.
It significantly reduces energy consumption during both training and inference.
arXiv Detail & Related papers (2024-05-25T19:50:23Z) - Enabling On-device Continual Learning with Binary Neural Networks [3.180732240499359]
We propose a solution that combines recent advancements in the field of Continual Learning (CL) and Binary Neural Networks (BNNs)
Specifically, our approach leverages binary latent replay activations and a novel quantization scheme that significantly reduces the number of bits required for gradient computation.
arXiv Detail & Related papers (2024-01-18T11:57:05Z) - Biologically Plausible Learning on Neuromorphic Hardware Architectures [27.138481022472]
Neuromorphic computing is an emerging paradigm that confronts this imbalance by computations directly in analog memories.
This work is the first to compare the impact of different learning algorithms on Compute-In-Memory-based hardware and vice versa.
arXiv Detail & Related papers (2022-12-29T15:10:59Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - A Stable, Fast, and Fully Automatic Learning Algorithm for Predictive
Coding Networks [65.34977803841007]
Predictive coding networks are neuroscience-inspired models with roots in both Bayesian statistics and neuroscience.
We show how by simply changing the temporal scheduling of the update rule for the synaptic weights leads to an algorithm that is much more efficient and stable than the original one.
arXiv Detail & Related papers (2022-11-16T00:11:04Z) - Improved Algorithms for Neural Active Learning [74.89097665112621]
We improve the theoretical and empirical performance of neural-network(NN)-based active learning algorithms for the non-parametric streaming setting.
We introduce two regret metrics by minimizing the population loss that are more suitable in active learning than the one used in state-of-the-art (SOTA) related work.
arXiv Detail & Related papers (2022-10-02T05:03:38Z) - Recognizing and Verifying Mathematical Equations using Multiplicative
Differential Neural Units [86.9207811656179]
We show that memory-augmented neural networks (NNs) can achieve higher-order, memory-augmented extrapolation, stable performance, and faster convergence.
Our models achieve a 1.53% average improvement over current state-of-the-art methods in equation verification and achieve a 2.22% Top-1 average accuracy and 2.96% Top-5 average accuracy for equation completion.
arXiv Detail & Related papers (2021-04-07T03:50:11Z) - Truly Sparse Neural Networks at Scale [2.2860412844991655]
We train the largest neural network ever trained in terms of representational power -- reaching the bat brain size.
Our approach has state-of-the-art performance while opening the path for an environmentally friendly artificial intelligence era.
arXiv Detail & Related papers (2021-02-02T20:06:47Z) - A Novel Anomaly Detection Algorithm for Hybrid Production Systems based
on Deep Learning and Timed Automata [73.38551379469533]
DAD:DeepAnomalyDetection is a new approach for automatic model learning and anomaly detection in hybrid production systems.
It combines deep learning and timed automata for creating behavioral model from observations.
The algorithm has been applied to few data sets including two from real systems and has shown promising results.
arXiv Detail & Related papers (2020-10-29T08:27:43Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.