Robust High-dimensional Memory-augmented Neural Networks
- URL: http://arxiv.org/abs/2010.01939v2
- Date: Fri, 19 Mar 2021 09:46:59 GMT
- Title: Robust High-dimensional Memory-augmented Neural Networks
- Authors: Geethan Karunaratne, Manuel Schmuck, Manuel Le Gallo, Giovanni
Cherubini, Luca Benini, Abu Sebastian, Abbas Rahimi
- Abstract summary: Memory-augmented neural networks enhance neural networks with an explicit memory to overcome these issues.
Access to this explicit memory occurs via soft read and write operations involving every individual memory entry.
We propose a robust architecture that employs a computational memory unit as the explicit memory performing analog in-memory computation on high-dimensional (HD) vectors.
- Score: 13.82206983716435
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traditional neural networks require enormous amounts of data to build their
complex mappings during a slow training procedure that hinders their abilities
for relearning and adapting to new data. Memory-augmented neural networks
enhance neural networks with an explicit memory to overcome these issues.
Access to this explicit memory, however, occurs via soft read and write
operations involving every individual memory entry, resulting in a bottleneck
when implemented using the conventional von Neumann computer architecture. To
overcome this bottleneck, we propose a robust architecture that employs a
computational memory unit as the explicit memory performing analog in-memory
computation on high-dimensional (HD) vectors, while closely matching 32-bit
software-equivalent accuracy. This is achieved by a content-based attention
mechanism that represents unrelated items in the computational memory with
uncorrelated HD vectors, whose real-valued components can be readily
approximated by binary, or bipolar components. Experimental results demonstrate
the efficacy of our approach on few-shot image classification tasks on the
Omniglot dataset using more than 256,000 phase-change memory devices. Our
approach effectively merges the richness of deep neural network representations
with HD computing that paves the way for robust vector-symbolic manipulations
applicable in reasoning, fusion, and compression.
Related papers
- Dynamic neural network with memristive CIM and CAM for 2D and 3D vision [57.6208980140268]
We propose a semantic memory-based dynamic neural network (DNN) using memristor.
The network associates incoming data with the past experience stored as semantic vectors.
We validate our co-designs, using a 40nm memristor macro, on ResNet and PointNet++ for classifying images and 3D points from the MNIST and ModelNet datasets.
arXiv Detail & Related papers (2024-07-12T04:55:57Z) - Efficient and accurate neural field reconstruction using resistive memory [52.68088466453264]
Traditional signal reconstruction methods on digital computers face both software and hardware challenges.
We propose a systematic approach with software-hardware co-optimizations for signal reconstruction from sparse inputs.
This work advances the AI-driven signal restoration technology and paves the way for future efficient and robust medical AI and 3D vision applications.
arXiv Detail & Related papers (2024-04-15T09:33:09Z) - Resistive Memory-based Neural Differential Equation Solver for Score-based Diffusion Model [55.116403765330084]
Current AIGC methods, such as score-based diffusion, are still deficient in terms of rapidity and efficiency.
We propose a time-continuous and analog in-memory neural differential equation solver for score-based diffusion.
We experimentally validate our solution with 180 nm resistive memory in-memory computing macros.
arXiv Detail & Related papers (2024-04-08T16:34:35Z) - Heterogenous Memory Augmented Neural Networks [84.29338268789684]
We introduce a novel heterogeneous memory augmentation approach for neural networks.
By introducing learnable memory tokens with attention mechanism, we can effectively boost performance without huge computational overhead.
We show our approach on various image and graph-based tasks under both in-distribution (ID) and out-of-distribution (OOD) conditions.
arXiv Detail & Related papers (2023-10-17T01:05:28Z) - A Co-design view of Compute in-Memory with Non-Volatile Elements for
Neural Networks [12.042322495445196]
We discuss how compute-in-memory can play an important part in the next generation of computing hardware.
A non-volatile memory based cross-bar architecture forms the heart of an engine that uses an analog process to parallelize the matrix vector multiplication operation.
The cross-bar architecture, at times referred to as a neuromorphic approach, can be a key hardware element in future computing machines.
arXiv Detail & Related papers (2022-06-03T15:59:46Z) - Experimentally realized memristive memory augmented neural network [0.0]
Lifelong on-device learning is a key challenge for machine intelligence.
Memory augmented neural network has been proposed to achieve the goal, but the memory module has to be stored in an off-chip memory.
We implement the entire memory augmented neural network architecture in a fully integrated memristive crossbar platform.
arXiv Detail & Related papers (2022-04-15T11:52:30Z) - MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning [72.80896338009579]
We find that the memory bottleneck is due to the imbalanced memory distribution in convolutional neural network (CNN) designs.
We propose a generic patch-by-patch inference scheduling, which significantly cuts down the peak memory.
We automate the process with neural architecture search to jointly optimize the neural architecture and inference scheduling, leading to MCUNetV2.
arXiv Detail & Related papers (2021-10-28T17:58:45Z) - Reservoir Memory Machines as Neural Computers [70.5993855765376]
Differentiable neural computers extend artificial neural networks with an explicit memory without interference.
We achieve some of the computational capabilities of differentiable neural computers with a model that can be trained very efficiently.
arXiv Detail & Related papers (2020-09-14T12:01:30Z) - Improving Memory Utilization in Convolutional Neural Network
Accelerators [16.340620299847384]
We propose a mapping method that allows activation layers to overlap and thus utilize the memory more efficiently.
Experiments with various real-world object detector networks show that the proposed mapping technique can decrease the activations memory by up to 32.9%.
For higher resolution de-noising networks, we achieve activation memory savings of 48.8%.
arXiv Detail & Related papers (2020-07-20T09:34:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.