Model of the Weak Reset Process in HfOx Resistive Memory for Deep
Learning Frameworks
- URL: http://arxiv.org/abs/2107.06064v1
- Date: Fri, 2 Jul 2021 08:50:35 GMT
- Title: Model of the Weak Reset Process in HfOx Resistive Memory for Deep
Learning Frameworks
- Authors: Atreya Majumdar, Marc Bocquet, Tifenn Hirtzlin, Axel Laborieux,
Jacques-Olivier Klein, Etienne Nowak, Elisa Vianello, Jean-Michel Portal,
Damien Querlioz
- Abstract summary: We present a model of the weak RESET process in hafnium oxide RRAM.
We integrate this model within the PyTorch deep learning framework.
We use this tool to train Binarized Neural Networks for the MNIST handwritten digit recognition task.
- Score: 0.6745502291821955
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The implementation of current deep learning training algorithms is
power-hungry, owing to data transfer between memory and logic units.
Oxide-based RRAMs are outstanding candidates to implement in-memory computing,
which is less power-intensive. Their weak RESET regime, is particularly
attractive for learning, as it allows tuning the resistance of the devices with
remarkable endurance. However, the resistive change behavior in this regime
suffers many fluctuations and is particularly challenging to model, especially
in a way compatible with tools used for simulating deep learning. In this work,
we present a model of the weak RESET process in hafnium oxide RRAM and
integrate this model within the PyTorch deep learning framework. Validated on
experiments on a hybrid CMOS/RRAM technology, our model reproduces both the
noisy progressive behavior and the device-to-device (D2D) variability. We use
this tool to train Binarized Neural Networks for the MNIST handwritten digit
recognition task and the CIFAR-10 object classification task. We simulate our
model with and without various aspects of device imperfections to understand
their impact on the training process and identify that the D2D variability is
the most detrimental aspect. The framework can be used in the same manner for
other types of memories to identify the device imperfections that cause the
most degradation, which can, in turn, be used to optimize the devices to reduce
the impact of these imperfections.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - Diffusion-Based Neural Network Weights Generation [80.89706112736353]
D2NWG is a diffusion-based neural network weights generation technique that efficiently produces high-performing weights for transfer learning.
Our method extends generative hyper-representation learning to recast the latent diffusion paradigm for neural network weights generation.
Our approach is scalable to large architectures such as large language models (LLMs), overcoming the limitations of current parameter generation techniques.
arXiv Detail & Related papers (2024-02-28T08:34:23Z) - Random resistive memory-based deep extreme point learning machine for
unified visual processing [67.51600474104171]
We propose a novel hardware-software co-design, random resistive memory-based deep extreme point learning machine (DEPLM)
Our co-design system achieves huge energy efficiency improvements and training cost reduction when compared to conventional systems.
arXiv Detail & Related papers (2023-12-14T09:46:16Z) - Synaptic metaplasticity with multi-level memristive devices [1.5598974049838272]
We propose a memristor-based hardware solution for implementing metaplasticity during both inference and training.
We show that a two-layer perceptron achieves 97% and 86% accuracy on consecutive training of MNIST and Fashion-MNIST.
Our architecture is compatible with the memristor limited endurance and has a 15x reduction in memory.
arXiv Detail & Related papers (2023-06-21T09:40:25Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - Real-to-Sim: Predicting Residual Errors of Robotic Systems with Sparse
Data using a Learning-based Unscented Kalman Filter [65.93205328894608]
We learn the residual errors between a dynamic and/or simulator model and the real robot.
We show that with the learned residual errors, we can further close the reality gap between dynamic models, simulations, and actual hardware.
arXiv Detail & Related papers (2022-09-07T15:15:12Z) - Incremental Online Learning Algorithms Comparison for Gesture and Visual
Smart Sensors [68.8204255655161]
This paper compares four state-of-the-art algorithms in two real applications: gesture recognition based on accelerometer data and image classification.
Our results confirm these systems' reliability and the feasibility of deploying them in tiny-memory MCUs.
arXiv Detail & Related papers (2022-09-01T17:05:20Z) - Online Training of Spiking Recurrent Neural Networks with Phase-Change
Memory Synapses [1.9809266426888898]
Training spiking neural networks (RNNs) on dedicated neuromorphic hardware is still an open challenge.
We present a simulation framework of differential-architecture arrays based on an accurate and comprehensive Phase-Change Memory (PCM) device model.
We train a spiking RNN whose weights are emulated in the presented simulation framework, using a recently proposed e-prop learning rule.
arXiv Detail & Related papers (2021-08-04T01:24:17Z) - Tiny Machine Learning for Concept Drift [8.452237741722726]
This paper introduces a Tiny Machine Learning for Concept Drift (TML-CD) solution based on deep learning feature extractors and a k-nearest neighbors.
The adaptation module continuously updates the knowledge base of TML-CD to deal with concept drift affecting the data-generating process.
The porting of TML-CD on three off-the-shelf micro-controller units shows the feasibility of what is proposed in real-world pervasive systems.
arXiv Detail & Related papers (2021-07-30T17:02:04Z) - Neuromodulated Neural Architectures with Local Error Signals for
Memory-Constrained Online Continual Learning [4.2903672492917755]
We develop a biologically-inspired light weight neural network architecture that incorporates local learning and neuromodulation.
We demonstrate the efficacy of our approach on both single task and continual learning setting.
arXiv Detail & Related papers (2020-07-16T07:41:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.