Towards a Scalable and Distributed Infrastructure for Deep Learning
Applications
- URL: http://arxiv.org/abs/2010.03012v2
- Date: Tue, 20 Apr 2021 00:18:24 GMT
- Title: Towards a Scalable and Distributed Infrastructure for Deep Learning
Applications
- Authors: Bita Hasheminezhad, Shahrzad Shirzad, Nanmiao Wu, Patrick Diehl,
Hannes Schulz, Hartmut Kaiser
- Abstract summary: Phylanx offers a productivity-oriented execution tree that can be executed on multiple nodes.
We present Phylanx that has the potential to alleviate shortcomings in distributed deep learning frameworks.
- Score: 4.4979162962108905
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although recent scaling up approaches to training deep neural networks have
proven to be effective, the computational intensity of large and complex
models, as well as the availability of large-scale datasets, require deep
learning frameworks to utilize scaling out techniques. Parallelization
approaches and distribution requirements are not considered in the preliminary
designs of most available distributed deep learning frameworks, and most of
them still are not able to perform effective and efficient fine-grained
inter-node communication. We present Phylanx that has the potential to
alleviate these shortcomings. Phylanx offers a productivity-oriented frontend
where user Python code is translated to a futurized execution tree that can be
executed efficiently on multiple nodes using the C++ standard library for
parallelism and concurrency (HPX), leveraging fine-grained threading and an
active messaging task-based runtime system.
Related papers
- Towards a Better Theoretical Understanding of Independent Subnetwork Training [56.24689348875711]
We take a closer theoretical look at Independent Subnetwork Training (IST)
IST is a recently proposed and highly effective technique for solving the aforementioned problems.
We identify fundamental differences between IST and alternative approaches, such as distributed methods with compressed communication.
arXiv Detail & Related papers (2023-06-28T18:14:22Z) - Partitioning Distributed Compute Jobs with Reinforcement Learning and
Graph Neural Networks [58.720142291102135]
Large-scale machine learning models are bringing advances to a broad range of fields.
Many of these models are too large to be trained on a single machine, and must be distributed across multiple devices.
We show that maximum parallelisation is sub-optimal in relation to user-critical metrics such as throughput and blocking rate.
arXiv Detail & Related papers (2023-01-31T17:41:07Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Nebula-I: A General Framework for Collaboratively Training Deep Learning
Models on Low-Bandwidth Cloud Clusters [39.85470606966918]
We introduce a general framework, Nebula-I, for collaboratively training deep learning models over remote heterogeneous clusters.
Nebula-I is implemented with the PaddlePaddle deep learning framework, which can support collaborative training over heterogeneous hardware.
Experiments demonstrate that the proposed framework could substantially maximize the training efficiency while preserving satisfactory NLP performance.
arXiv Detail & Related papers (2022-05-19T11:10:14Z) - OneFlow: Redesign the Distributed Deep Learning Framework from Scratch [17.798586916628174]
OneFlow is a novel distributed training framework based on an SBP (split, broadcast and partial-value) abstraction and the actor model.
SBP enables much easier programming of data parallelism and model parallelism than existing frameworks.
OneFlow outperforms many well-known customized libraries built on top of the state-of-the-art frameworks.
arXiv Detail & Related papers (2021-10-28T11:32:14Z) - Parallel Training of Deep Networks with Local Updates [84.30918922367442]
Local parallelism is a framework which parallelizes training of individual layers in deep networks by replacing global backpropagation with truncated layer-wise backpropagation.
We show results in both vision and language domains across a diverse set of architectures, and find that local parallelism is particularly effective in the high-compute regime.
arXiv Detail & Related papers (2020-12-07T16:38:45Z) - Benchmarking network fabrics for data distributed training of deep
neural networks [10.067102343753643]
Large computational requirements for training deep models have necessitated the development of new methods for faster training.
One such approach is the data parallel approach, where the training data is distributed across multiple compute nodes.
In this paper, we examine the effects of using different physical hardware interconnects and network-related software primitives for enabling data distributed deep learning.
arXiv Detail & Related papers (2020-08-18T17:38:30Z) - Einsum Networks: Fast and Scalable Learning of Tractable Probabilistic
Circuits [99.59941892183454]
We propose Einsum Networks (EiNets), a novel implementation design for PCs.
At their core, EiNets combine a large number of arithmetic operations in a single monolithic einsum-operation.
We show that the implementation of Expectation-Maximization (EM) can be simplified for PCs, by leveraging automatic differentiation.
arXiv Detail & Related papers (2020-04-13T23:09:15Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.