Replicate or Relocate? Non-Uniform Access in Parameter Servers
- URL: http://arxiv.org/abs/2104.00501v1
- Date: Thu, 1 Apr 2021 14:52:32 GMT
- Title: Replicate or Relocate? Non-Uniform Access in Parameter Servers
- Authors: Alexander Renz-Wieland, Rainer Gemulla, Zoi Kaoudi, Volker Markl
- Abstract summary: We present Lapse2, a PS that replicates hot spot parameters, relocates less frequently accessed parameters, and employs specialized techniques to manage nondeterminism.
In our experimental study, Lapse2 outperformed existing, single-technique PSs by up to one order of magnitude.
- Score: 74.89066750738971
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Parameter servers (PSs) facilitate the implementation of distributed training
for large machine learning tasks. A key challenge for PS performance is that
parameter access is non-uniform in many real-world machine learning tasks,
i.e., different parameters exhibit drastically different access patterns. We
identify skew and nondeterminism as two major sources for non-uniformity.
Existing PSs are ill-suited for managing such non-uniform access because they
uniformly apply the same parameter management technique to all parameters. As
consequence, the performance of existing PSs is negatively affected and may
even fall behind that of single node baselines. In this paper, we explore how
PSs can manage non-uniform access efficiently. We find that it is key for PSs
to support multiple management techniques and to leverage a well-suited
management technique for each parameter. We present Lapse2, a PS that
replicates hot spot parameters, relocates less frequently accessed parameters,
and employs specialized techniques to manage nondeterminism that arises from
random sampling. In our experimental study, Lapse2 outperformed existing,
single-technique PSs by up to one order of magnitude and provided near-linear
scalability across multiple machine learning tasks.
Related papers
- Sparse Diffusion Policy: A Sparse, Reusable, and Flexible Policy for Robot Learning [61.294110816231886]
We introduce a sparse, reusable, and flexible policy, Sparse Diffusion Policy (SDP)
SDP selectively activates experts and skills, enabling efficient and task-specific learning without retraining the entire model.
Demos and codes can be found in https://forrest-110.io/sparse_diffusion_policy/.
arXiv Detail & Related papers (2024-07-01T17:59:56Z) - DiSparse: Disentangled Sparsification for Multitask Model Compression [92.84435347164435]
DiSparse is a simple, effective, and first-of-its-kind multitask pruning and sparse training scheme.
Our experimental results demonstrate superior performance on various configurations and settings.
arXiv Detail & Related papers (2022-06-09T17:57:46Z) - Good Intentions: Adaptive Parameter Management via Intent Signaling [50.01012642343155]
We propose a novel intent signaling mechanism that integrates naturally into existing machine learning stacks.
We then describe AdaPM, a fully adaptive, zero-tuning parameter manager based on this mechanism.
In our evaluation, AdaPM matched or outperformed state-of-the-art parameter managers out of the box.
arXiv Detail & Related papers (2022-06-01T13:02:19Z) - Parameter-Efficient Transfer Learning with Diff Pruning [108.03864629388404]
diff pruning is a simple approach to enable parameter-efficient transfer learning within the pretrain-finetune framework.
We find that models finetuned with diff pruning can match the performance of fully finetuned baselines on the GLUE benchmark.
arXiv Detail & Related papers (2020-12-14T12:34:01Z) - Dynamic Parameter Allocation in Parameter Servers [74.250687861348]
We propose to integrate dynamic parameter allocation into parameter servers, describe an efficient implementation of such a parameter server called Lapse.
We found that Lapse provides near-linear scaling and can be orders of magnitude faster than existing parameter servers.
arXiv Detail & Related papers (2020-02-03T11:37:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.