Learning Without a Global Clock: Asynchronous Learning in a
Physics-Driven Learning Network
- URL: http://arxiv.org/abs/2201.04626v1
- Date: Mon, 10 Jan 2022 05:38:01 GMT
- Title: Learning Without a Global Clock: Asynchronous Learning in a
Physics-Driven Learning Network
- Authors: Jacob F Wycoff, Sam Dillavou, Menachem Stern, Andrea J Liu, Douglas J
Durian
- Abstract summary: We show that desynchronizing the learning process does not degrade performance for a variety of tasks in an idealized simulation.
We draw an analogy between asynchronicity and mini-batching in gradient descent, and show that they have similar effects on the learning process.
- Score: 1.3124513975412255
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In a neuron network, synapses update individually using local information,
allowing for entirely decentralized learning. In contrast, elements in an
artificial neural network (ANN) are typically updated simultaneously using a
central processor. Here we investigate the feasibility and effect of
asynchronous learning in a recently introduced decentralized, physics-driven
learning network. We show that desynchronizing the learning process does not
degrade performance for a variety of tasks in an idealized simulation. In
experiment, desynchronization actually improves performance by allowing the
system to better explore the discretized state space of solutions. We draw an
analogy between asynchronicity and mini-batching in stochastic gradient
descent, and show that they have similar effects on the learning process.
Desynchronizing the learning process establishes physics-driven learning
networks as truly fully distributed learning machines, promoting better
performance and scalability in deployment.
Related papers
- A Unified Framework for Neural Computation and Learning Over Time [56.44910327178975]
Hamiltonian Learning is a novel unified framework for learning with neural networks "over time"
It is based on differential equations that: (i) can be integrated without the need of external software solvers; (ii) generalize the well-established notion of gradient-based learning in feed-forward and recurrent networks; (iii) open to novel perspectives.
arXiv Detail & Related papers (2024-09-18T14:57:13Z) - Overcoming the Limitations of Layer Synchronization in Spiking Neural Networks [0.11522790873450185]
A truly asynchronous system would allow all neurons to evaluate concurrently their threshold and emit spikes upon receiving any presynaptic current.
We present a study that documents and quantifies this problem in three datasets on our simulation environment that implements network asynchrony.
We show that models trained with layer synchronization either perform sub-optimally in absence of the synchronization, or they will fail to benefit from any energy and latency reduction.
arXiv Detail & Related papers (2024-08-09T14:39:23Z) - Neuromorphic dreaming: A pathway to efficient learning in artificial agents [2.6542148964152923]
We present a hardware implementation of model-based reinforcement learning (MBRL) using spiking neural networks (SNNs) on mixed-signal analog/digital neuromorphic hardware.
This approach leverages the energy efficiency of mixed-signal neuromorphic chips while achieving high sample efficiency.
We validate the model by training the hardware implementation to play the Atari game Pong.
arXiv Detail & Related papers (2024-05-24T15:03:56Z) - Stragglers-Aware Low-Latency Synchronous Federated Learning via Layer-Wise Model Updates [71.81037644563217]
Synchronous federated learning (FL) is a popular paradigm for collaborative edge learning.
As some of the devices may have limited computational resources and varying availability, FL latency is highly sensitive to stragglers.
We propose straggler-aware layer-wise federated learning (SALF) that leverages the optimization procedure of NNs via backpropagation to update the global model in a layer-wise fashion.
arXiv Detail & Related papers (2024-03-27T09:14:36Z) - An effective and efficient green federated learning method for one-layer
neural networks [0.22499166814992436]
Federated learning (FL) is one of the most active research lines in machine learning.
We present a FL method, based on a neural network without hidden layers, capable of generating a global collaborative model in a single training round.
We show that the method performs equally well in both identically and non-identically distributed scenarios.
arXiv Detail & Related papers (2023-12-22T08:52:08Z) - Asynchronous Reinforcement Learning for Real-Time Control of Physical
Robots [2.3061446605472558]
We show that when learning updates are expensive, the performance of sequential learning diminishes and is outperformed by asynchronous learning by a substantial margin.
Our system learns in real-time to reach and track visual targets from pixels within two hours of experience and does so directly using real robots.
arXiv Detail & Related papers (2022-03-23T23:05:28Z) - Characterizing and overcoming the greedy nature of learning in
multi-modal deep neural networks [62.48782506095565]
We show that due to the greedy nature of learning in deep neural networks, models tend to rely on just one modality while under-fitting the other modalities.
We propose an algorithm to balance the conditional learning speeds between modalities during training and demonstrate that it indeed addresses the issue of greedy learning.
arXiv Detail & Related papers (2022-02-10T20:11:21Z) - Accelerating Neural Network Training with Distributed Asynchronous and
Selective Optimization (DASO) [0.0]
We introduce the Distributed Asynchronous and Selective Optimization (DASO) method to accelerate network training.
DASO uses a hierarchical and asynchronous communication scheme comprised of node-local and global networks.
We show that DASO yields a reduction in training time of up to 34% on classical and state-of-the-art networks.
arXiv Detail & Related papers (2021-04-12T16:02:20Z) - Unsupervised Learning for Asynchronous Resource Allocation in Ad-hoc
Wireless Networks [122.42812336946756]
We design an unsupervised learning method based on Aggregation Graph Neural Networks (Agg-GNNs)
We capture the asynchrony by modeling the activation pattern as a characteristic of each node and train a policy-based resource allocation method.
arXiv Detail & Related papers (2020-11-05T03:38:36Z) - A Low Complexity Decentralized Neural Net with Centralized Equivalence
using Layer-wise Learning [49.15799302636519]
We design a low complexity decentralized learning algorithm to train a recently proposed large neural network in distributed processing nodes (workers)
In our setup, the training data is distributed among the workers but is not shared in the training process due to privacy and security concerns.
We show that it is possible to achieve equivalent learning performance as if the data is available in a single place.
arXiv Detail & Related papers (2020-09-29T13:08:12Z) - The large learning rate phase of deep learning: the catapult mechanism [50.23041928811575]
We present a class of neural networks with solvable training dynamics.
We find good agreement between our model's predictions and training dynamics in realistic deep learning settings.
We believe our results shed light on characteristics of models trained at different learning rates.
arXiv Detail & Related papers (2020-03-04T17:52:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.