PARTIME: Scalable and Parallel Processing Over Time with Deep Neural
Networks
- URL: http://arxiv.org/abs/2210.09147v1
- Date: Mon, 17 Oct 2022 14:49:14 GMT
- Title: PARTIME: Scalable and Parallel Processing Over Time with Deep Neural
Networks
- Authors: Enrico Meloni, Lapo Faggi, Simone Marullo, Alessandro Betti, Matteo
Tiezzi, Marco Gori, Stefano Melacci
- Abstract summary: We present PARTIME, a library designed to speed up neural networks whenever data is continuously streamed over time.
PARTIME starts processing each data sample at the time in which it becomes available from the stream.
Experiments are performed in order to empirically compare PARTIME with classic non-parallel neural computations in online learning.
- Score: 68.96484488899901
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we present PARTIME, a software library written in Python and
based on PyTorch, designed specifically to speed up neural networks whenever
data is continuously streamed over time, for both learning and inference.
Existing libraries are designed to exploit data-level parallelism, assuming
that samples are batched, a condition that is not naturally met in applications
that are based on streamed data. Differently, PARTIME starts processing each
data sample at the time in which it becomes available from the stream. PARTIME
wraps the code that implements a feed-forward multi-layer network and it
distributes the layer-wise processing among multiple devices, such as Graphics
Processing Units (GPUs). Thanks to its pipeline-based computational scheme,
PARTIME allows the devices to perform computations in parallel. At inference
time this results in scaling capabilities that are theoretically linear with
respect to the number of devices. During the learning stage, PARTIME can
leverage the non-i.i.d. nature of the streamed data with samples that are
smoothly evolving over time for efficient gradient computations. Experiments
are performed in order to empirically compare PARTIME with classic non-parallel
neural computations in online learning, distributing operations on up to 8
NVIDIA GPUs, showing significant speedups that are almost linear in the number
of devices, mitigating the impact of the data transfer overhead.
Related papers
- Streaming Factor Trajectory Learning for Temporal Tensor Decomposition [33.18423605559094]
We propose Streaming Factor Trajectory Learning for temporal tensor decomposition.
We use Gaussian processes (GPs) to model the trajectory of factors so as to flexibly estimate their temporal evolution.
We have shown the advantage of SFTL in both synthetic tasks and real-world applications.
arXiv Detail & Related papers (2023-10-25T21:58:52Z) - Tensor Slicing and Optimization for Multicore NPUs [2.670309629218727]
This paper proposes a compiler optimization pass for Multicore NPUs, called Slicing Optimization (TSO)
TSO identifies the best tensor slicing that minimizes execution time for a set of CNN models.
Results show that TSO is capable of identifying the best tensor slicing that minimizes execution time for a set of CNN models.
arXiv Detail & Related papers (2023-04-06T12:03:03Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Accelerating GAN training using highly parallel hardware on public cloud [0.3694429692322631]
This work explores different types of cloud services to train a Geneversarative Adversarial Network (GAN) in a parallel environment.
We parallelize the training process on multiple GPU and Google Processing Units (TPU)
Linear speed-up of the training process is obtained, while retaining most of the performance in terms of physics results.
arXiv Detail & Related papers (2021-11-08T16:59:15Z) - Accelerating Training and Inference of Graph Neural Networks with Fast
Sampling and Pipelining [58.10436813430554]
Mini-batch training of graph neural networks (GNNs) requires a lot of computation and data movement.
We argue in favor of performing mini-batch training with neighborhood sampling in a distributed multi-GPU environment.
We present a sequence of improvements to mitigate these bottlenecks, including a performance-engineered neighborhood sampler.
We also conduct an empirical analysis that supports the use of sampling for inference, showing that test accuracies are not materially compromised.
arXiv Detail & Related papers (2021-10-16T02:41:35Z) - OpTorch: Optimized deep learning architectures for resource limited
environments [1.5736899098702972]
We propose optimized deep learning pipelines in multiple aspects of training including time and memory.
OpTorch is a machine learning library designed to overcome weaknesses in existing implementations of neural network training.
arXiv Detail & Related papers (2021-05-03T03:58:57Z) - Parareal Neural Networks Emulating a Parallel-in-time Algorithm [1.988145627448243]
As deep neural networks (DNNs) become deeper, the training time increases.
In this paper, we introduce a novel methodology to construct a parallel neural network.
arXiv Detail & Related papers (2021-03-16T02:03:39Z) - Deep Cellular Recurrent Network for Efficient Analysis of Time-Series
Data with Spatial Information [52.635997570873194]
This work proposes a novel deep cellular recurrent neural network (DCRNN) architecture to process complex multi-dimensional time series data with spatial information.
The proposed architecture achieves state-of-the-art performance while utilizing substantially less trainable parameters when compared to comparable methods in the literature.
arXiv Detail & Related papers (2021-01-12T20:08:18Z) - Unsupervised Parallel Corpus Mining on Web Data [53.74427402568838]
We present a pipeline to mine the parallel corpus from the Internet in an unsupervised manner.
Our system produces new state-of-the-art results, 39.81 and 38.95 BLEU scores, even compared with supervised approaches.
arXiv Detail & Related papers (2020-09-18T02:38:01Z) - Accelerating Feedforward Computation via Parallel Nonlinear Equation
Solving [106.63673243937492]
Feedforward computation, such as evaluating a neural network or sampling from an autoregressive model, is ubiquitous in machine learning.
We frame the task of feedforward computation as solving a system of nonlinear equations. We then propose to find the solution using a Jacobi or Gauss-Seidel fixed-point method, as well as hybrid methods of both.
Our method is guaranteed to give exactly the same values as the original feedforward computation with a reduced (or equal) number of parallelizable iterations, and hence reduced time given sufficient parallel computing power.
arXiv Detail & Related papers (2020-02-10T10:11:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.