Accelerating Hierarchical Associative Memory: A Deep Equilibrium
Approach
- URL: http://arxiv.org/abs/2311.15673v1
- Date: Mon, 27 Nov 2023 10:02:12 GMT
- Title: Accelerating Hierarchical Associative Memory: A Deep Equilibrium
Approach
- Authors: C\'edric Goemaere, Johannes Deleu, Thomas Demeester
- Abstract summary: We propose two strategies to speed up memory retrieval in Hierarchical Associative Memory models.
First, we show how they can be cast as Deep Equilibrium Models, which allows using faster and more stable solvers.
Second, inspired by earlier work, we show that alternating optimization of the even and odd layers accelerates memory retrieval by a factor close to two.
- Score: 12.829893293085732
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Hierarchical Associative Memory models have recently been proposed as a
versatile extension of continuous Hopfield networks. In order to facilitate
future research on such models, especially at scale, we focus on increasing
their simulation efficiency on digital hardware. In particular, we propose two
strategies to speed up memory retrieval in these models, which corresponds to
their use at inference, but is equally important during training. First, we
show how they can be cast as Deep Equilibrium Models, which allows using faster
and more stable solvers. Second, inspired by earlier work, we show that
alternating optimization of the even and odd layers accelerates memory
retrieval by a factor close to two. Combined, these two techniques allow for a
much faster energy minimization, as shown in our proof-of-concept experimental
results. The code is available at https://github.com/cgoemaere/hamdeq
Related papers
- KITINet: Kinetics Theory Inspired Network Architectures with PDE Simulation Approaches [43.872190335490515]
This paper introduces KITINet, a novel architecture that reinterprets feature propagation through the lens of non-equilibrium particle dynamics.<n>At its core, we propose a residual module that models update as the evolution of a particle system.<n>This formulation mimics particle collisions and energy exchange, enabling adaptive feature refinement via physics-informed interactions.
arXiv Detail & Related papers (2025-05-23T13:58:29Z) - Efficient Diffusion Training through Parallelization with Truncated Karhunen-Loève Expansion [5.770347328961063]
Diffusion denoising models suffer from slow convergence during training.
We propose a novel forward-time process for training and sampling.
Our method significantly outperforms baseline diffusion models.
arXiv Detail & Related papers (2025-03-22T05:34:02Z) - Neural Hamilton: Can A.I. Understand Hamiltonian Mechanics? [0.0]
We propose a novel framework based on neural network that reformulates classical mechanics as an operator learning problem.
A machine directly maps a potential function to its corresponding trajectory in phase space without solving the Hamilton equations.
arXiv Detail & Related papers (2024-10-28T12:10:44Z) - TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - An Optimization-based Deep Equilibrium Model for Hyperspectral Image
Deconvolution with Convergence Guarantees [71.57324258813675]
We propose a novel methodology for addressing the hyperspectral image deconvolution problem.
A new optimization problem is formulated, leveraging a learnable regularizer in the form of a neural network.
The derived iterative solver is then expressed as a fixed-point calculation problem within the Deep Equilibrium framework.
arXiv Detail & Related papers (2023-06-10T08:25:16Z) - Neural Basis Functions for Accelerating Solutions to High Mach Euler
Equations [63.8376359764052]
We propose an approach to solving partial differential equations (PDEs) using a set of neural networks.
We regress a set of neural networks onto a reduced order Proper Orthogonal Decomposition (POD) basis.
These networks are then used in combination with a branch network that ingests the parameters of the prescribed PDE to compute a reduced order approximation to the PDE.
arXiv Detail & Related papers (2022-08-02T18:27:13Z) - Discovering dynamical features of Hodgkin-Huxley-type model of
physiological neuron using artificial neural network [0.0]
We consider Hodgkin-Huxley-type system with two fast and one slow variables.
For these two systems we create artificial neural networks that are able to reproduce their dynamics.
For the bistable model it means that the network being trained only on one brunch of the solutions recovers another without seeing it during the training.
arXiv Detail & Related papers (2022-03-26T19:04:19Z) - Physics Informed RNN-DCT Networks for Time-Dependent Partial
Differential Equations [62.81701992551728]
We present a physics-informed framework for solving time-dependent partial differential equations.
Our model utilizes discrete cosine transforms to encode spatial and recurrent neural networks.
We show experimental results on the Taylor-Green vortex solution to the Navier-Stokes equations.
arXiv Detail & Related papers (2022-02-24T20:46:52Z) - Unified Field Theory for Deep and Recurrent Neural Networks [56.735884560668985]
We present a unified and systematic derivation of the mean-field theory for both recurrent and deep networks.
We find that convergence towards the mean-field theory is typically slower for recurrent networks than for deep networks.
Our method exposes that Gaussian processes are but the lowest order of a systematic expansion in $1/n$.
arXiv Detail & Related papers (2021-12-10T15:06:11Z) - Stabilizing Equilibrium Models by Jacobian Regularization [151.78151873928027]
Deep equilibrium networks (DEQs) are a new class of models that eschews traditional depth in favor of finding the fixed point of a single nonlinear layer.
We propose a regularization scheme for DEQ models that explicitly regularizes the Jacobian of the fixed-point update equations to stabilize the learning of equilibrium models.
We show that this regularization adds only minimal computational cost, significantly stabilizes the fixed-point convergence in both forward and backward passes, and scales well to high-dimensional, realistic domains.
arXiv Detail & Related papers (2021-06-28T00:14:11Z) - Stragglers Are Not Disaster: A Hybrid Federated Learning Algorithm with
Delayed Gradients [21.63719641718363]
Federated learning (FL) is a new machine learning framework which trains a joint model across a large amount of decentralized computing devices.
This paper presents a novel FL algorithm, namely Hybrid Federated Learning (HFL), to achieve a learning balance in efficiency and effectiveness.
arXiv Detail & Related papers (2021-02-12T02:27:44Z) - Forecasting Sequential Data using Consistent Koopman Autoencoders [52.209416711500005]
A new class of physics-based methods related to Koopman theory has been introduced, offering an alternative for processing nonlinear dynamical systems.
We propose a novel Consistent Koopman Autoencoder model which, unlike the majority of existing work, leverages the forward and backward dynamics.
Key to our approach is a new analysis which explores the interplay between consistent dynamics and their associated Koopman operators.
arXiv Detail & Related papers (2020-03-04T18:24:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.