Improving Inference Lifetime of Neuromorphic Systems via Intelligent
Synapse Mapping
- URL: http://arxiv.org/abs/2106.09104v1
- Date: Wed, 16 Jun 2021 20:12:47 GMT
- Title: Improving Inference Lifetime of Neuromorphic Systems via Intelligent
Synapse Mapping
- Authors: Shihao Song, Twisha Titirsha, Anup Das
- Abstract summary: An RRAM cell can switch its state after reading its content a certain number of times.
We propose an architectural solution to extend the read endurance of RRAM-based neuromorphic systems.
- Score: 0.2578242050187029
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Non-Volatile Memories (NVMs) such as Resistive RAM (RRAM) are used in
neuromorphic systems to implement high-density and low-power analog synaptic
weights. Unfortunately, an RRAM cell can switch its state after reading its
content a certain number of times. Such behavior challenges the integrity and
program-once-read-many-times philosophy of implementing machine learning
inference on neuromorphic systems, impacting the Quality-of-Service (QoS).
Elevated temperatures and frequent usage can significantly shorten the number
of times an RRAM cell can be reliably read before it becomes absolutely
necessary to reprogram. We propose an architectural solution to extend the read
endurance of RRAM-based neuromorphic systems. We make two key contributions.
First, we formulate the read endurance of an RRAM cell as a function of the
programmed synaptic weight and its activation within a machine learning
workload. Second, we propose an intelligent workload mapping strategy
incorporating the endurance formulation to place the synapses of a machine
learning model onto the RRAM cells of the hardware. The objective is to extend
the inference lifetime, defined as the number of times the model can be used to
generate output (inference) before the trained weights need to be reprogrammed
on the RRAM cells of the system. We evaluate our architectural solution with
machine learning workloads on a cycle-accurate simulator of an RRAM-based
neuromorphic system. Our results demonstrate a significant increase in
inference lifetime with only a minimal performance impact.
Related papers
- DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution [114.61347672265076]
Development of MLLMs for real-world robots is challenging due to the typically limited computation and memory capacities available on robotic platforms.
We propose a Dynamic Early-Exit Framework for Robotic Vision-Language-Action Model (DeeR) that automatically adjusts the size of the activated MLLM.
DeeR demonstrates significant reductions in computational costs of LLM by 5.2-6.5x and GPU memory of LLM by 2-6x without compromising performance.
arXiv Detail & Related papers (2024-11-04T18:26:08Z) - A Realistic Simulation Framework for Analog/Digital Neuromorphic Architectures [73.65190161312555]
ARCANA is a spiking neural network simulator designed to account for the properties of mixed-signal neuromorphic circuits.
We show how the results obtained provide a reliable estimate of the behavior of the spiking neural network trained in software.
arXiv Detail & Related papers (2024-09-23T11:16:46Z) - Resistive Memory-based Neural Differential Equation Solver for Score-based Diffusion Model [55.116403765330084]
Current AIGC methods, such as score-based diffusion, are still deficient in terms of rapidity and efficiency.
We propose a time-continuous and analog in-memory neural differential equation solver for score-based diffusion.
We experimentally validate our solution with 180 nm resistive memory in-memory computing macros.
arXiv Detail & Related papers (2024-04-08T16:34:35Z) - Multi-level, Forming Free, Bulk Switching Trilayer RRAM for Neuromorphic
Computing at the Edge [0.0]
We develop a forming-free and bulk switching RRAM technology based on a trilayer metal-oxide stack.
We develop a neuromorphic compute-in-memory platform based on trilayer bulk RRAM crossbars.
Our work paves the way for neuromorphic computing at the edge under strict size, weight, and power constraints.
arXiv Detail & Related papers (2023-10-20T22:37:46Z) - Evaluation of STT-MRAM as a Scratchpad for Training in ML Accelerators [9.877596714655096]
Training deep neural networks (DNNs) is an extremely memory-intensive process.
Spin-Transfer-Torque MRAM (STT-MRAM) offers several desirable properties for training accelerators.
We show that MRAM provide up to 15-22x improvement in system level energy.
arXiv Detail & Related papers (2023-08-03T20:36:48Z) - The Expressive Leaky Memory Neuron: an Efficient and Expressive Phenomenological Neuron Model Can Solve Long-Horizon Tasks [64.08042492426992]
We introduce the Expressive Memory (ELM) neuron model, a biologically inspired model of a cortical neuron.
Our ELM neuron can accurately match the aforementioned input-output relationship with under ten thousand trainable parameters.
We evaluate it on various tasks with demanding temporal structures, including the Long Range Arena (LRA) datasets.
arXiv Detail & Related papers (2023-06-14T13:34:13Z) - Hardware calibrated learning to compensate heterogeneity in analog
RRAM-based Spiking Neural Networks [0.0]
Spiking Neural Networks (SNNs) can unleash the full power of analog Resistive Random Access Memories (RRAMs)
Their inherent computational sparsity naturally results in energy efficiency benefits.
Main challenge implementing robust SNNs is the intrinsic variability intrinsic (heterogeneity) of both analog CMOS circuits and RRAM technology.
arXiv Detail & Related papers (2022-02-10T15:33:03Z) - On the Mitigation of Read Disturbances in Neuromorphic Inference
Hardware [0.22940141855172028]
Non-Volatile Memory (NVM) cells are used in neuromorphic hardware to store model parameters.
NVM cells suffer from the read disturb issue, where the programmed resistance state drifts upon repeated access of a cell during inference.
We propose a system software framework to incorporate such dependencies in programming model parameters on NVM cells of a neuromorphic hardware.
arXiv Detail & Related papers (2022-01-27T14:02:54Z) - Mapping and Validating a Point Neuron Model on Intel's Neuromorphic
Hardware Loihi [77.34726150561087]
We investigate the potential of Intel's fifth generation neuromorphic chip - Loihi'
Loihi is based on the novel idea of Spiking Neural Networks (SNNs) emulating the neurons in the brain.
We find that Loihi replicates classical simulations very efficiently and scales notably well in terms of both time and energy performance as the networks get larger.
arXiv Detail & Related papers (2021-09-22T16:52:51Z) - Model of the Weak Reset Process in HfOx Resistive Memory for Deep
Learning Frameworks [0.6745502291821955]
We present a model of the weak RESET process in hafnium oxide RRAM.
We integrate this model within the PyTorch deep learning framework.
We use this tool to train Binarized Neural Networks for the MNIST handwritten digit recognition task.
arXiv Detail & Related papers (2021-07-02T08:50:35Z) - One-step regression and classification with crosspoint resistive memory
arrays [62.997667081978825]
High speed, low energy computing machines are in demand to enable real-time artificial intelligence at the edge.
One-step learning is supported by simulations of the prediction of the cost of a house in Boston and the training of a 2-layer neural network for MNIST digit recognition.
Results are all obtained in one computational step, thanks to the physical, parallel, and analog computing within the crosspoint array.
arXiv Detail & Related papers (2020-05-05T08:00:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.