Cost Effective MLaaS Federation: A Combinatorial Reinforcement Learning
Approach
- URL: http://arxiv.org/abs/2204.13971v1
- Date: Fri, 29 Apr 2022 09:44:04 GMT
- Title: Cost Effective MLaaS Federation: A Combinatorial Reinforcement Learning
Approach
- Authors: Shuzhao Xie, Yuan Xue, Yifei Zhu, and Zhi Wang
- Abstract summary: Federating different MLes together allows us to improve the analytic performance further.
naively aggregating results from different MLes not only incurs significant momentary cost but also may lead to sub-optimal performance gain.
We propose a framework fed Armol to unify the right selection of ML providers to achieve the best possible analytic performance.
- Score: 9.50492686145041
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the advancement of deep learning techniques, major cloud providers and
niche machine learning service providers start to offer their cloud-based
machine learning tools, also known as machine learning as a service (MLaaS), to
the public. According to our measurement, for the same task, these MLaaSes from
different providers have varying performance due to the proprietary datasets,
models, etc. Federating different MLaaSes together allows us to improve the
analytic performance further. However, naively aggregating results from
different MLaaSes not only incurs significant momentary cost but also may lead
to sub-optimal performance gain due to the introduction of possible
false-positive results. In this paper, we propose Armol, a framework to
federate the right selection of MLaaS providers to achieve the best possible
analytic performance. We first design a word grouping algorithm to unify the
output labels across different providers. We then present a deep combinatorial
reinforcement learning based-approach to maximize the accuracy while minimizing
the cost. The predictions from the selected providers are then aggregated
together using carefully chosen ensemble strategies. The real-world
trace-driven evaluation further demonstrates that Armol is able to achieve the
same accuracy results with $67\%$ less inference cost.
Related papers
- Routing to the Expert: Efficient Reward-guided Ensemble of Large
Language Models [69.51130760097818]
We propose Zooter, a reward-guided routing method distilling rewards on training queries to train a routing function.
We evaluate Zooter on a comprehensive benchmark collection with 26 subsets on different domains and tasks.
arXiv Detail & Related papers (2023-11-15T04:40:43Z) - FedLALR: Client-Specific Adaptive Learning Rates Achieve Linear Speedup
for Non-IID Data [54.81695390763957]
Federated learning is an emerging distributed machine learning method.
We propose a heterogeneous local variant of AMSGrad, named FedLALR, in which each client adjusts its learning rate.
We show that our client-specified auto-tuned learning rate scheduling can converge and achieve linear speedup with respect to the number of clients.
arXiv Detail & Related papers (2023-09-18T12:35:05Z) - FederatedScope-LLM: A Comprehensive Package for Fine-tuning Large
Language Models in Federated Learning [70.38817963253034]
This paper first discusses these challenges of federated fine-tuning LLMs, and introduces our package FS-LLM as a main contribution.
We provide comprehensive federated parameter-efficient fine-tuning algorithm implementations and versatile programming interfaces for future extension in FL scenarios.
We conduct extensive experiments to validate the effectiveness of FS-LLM and benchmark advanced LLMs with state-of-the-art parameter-efficient fine-tuning algorithms in FL settings.
arXiv Detail & Related papers (2023-09-01T09:40:36Z) - From Quantity to Quality: Boosting LLM Performance with Self-Guided Data Selection for Instruction Tuning [52.257422715393574]
We introduce a self-guided methodology for Large Language Models (LLMs) to autonomously discern and select cherry samples from open-source datasets.
Our key innovation, the Instruction-Following Difficulty (IFD) metric, emerges as a pivotal metric to identify discrepancies between a model's expected responses and its intrinsic generation capability.
arXiv Detail & Related papers (2023-08-23T09:45:29Z) - ezDPS: An Efficient and Zero-Knowledge Machine Learning Inference
Pipeline [2.0813318162800707]
We propose ezDPS, a new efficient and zero-knowledge Machine Learning inference scheme.
ezDPS is a zkML pipeline in which the data is processed in multiple stages for high accuracy.
We show that ezDPS achieves one-to-three orders of magnitude more efficient than the generic circuit-based approach in all metrics.
arXiv Detail & Related papers (2022-12-11T06:47:28Z) - Fed-CBS: A Heterogeneity-Aware Client Sampling Mechanism for Federated
Learning via Class-Imbalance Reduction [76.26710990597498]
We show that the class-imbalance of the grouped data from randomly selected clients can lead to significant performance degradation.
Based on our key observation, we design an efficient client sampling mechanism, i.e., Federated Class-balanced Sampling (Fed-CBS)
In particular, we propose a measure of class-imbalance and then employ homomorphic encryption to derive this measure in a privacy-preserving way.
arXiv Detail & Related papers (2022-09-30T05:42:56Z) - Active Learning at the ImageNet Scale [43.595076693347835]
In this work, we study a combination of active learning (AL) and pretraining (SSP) on ImageNet.
We find that performance on small toy datasets is not representative of performance on ImageNet due to the class imbalanced samples selected by an active learner.
We propose Balanced Selection (BASE), a simple, scalable AL algorithm that outperforms random sampling consistently.
arXiv Detail & Related papers (2021-11-25T02:48:51Z) - Towards Demystifying Serverless Machine Learning Training [19.061432528378788]
We present a systematic, comparative study of distributed machine learning training over serverless infrastructures.
We develop an analytic model to capture cost/performance tradeoffs that must be considered when opting for serverless infrastructure.
arXiv Detail & Related papers (2021-05-17T13:19:23Z) - Leveraging Expert Consistency to Improve Algorithmic Decision Support [62.61153549123407]
We explore the use of historical expert decisions as a rich source of information that can be combined with observed outcomes to narrow the construct gap.
We propose an influence function-based methodology to estimate expert consistency indirectly when each case in the data is assessed by a single expert.
Our empirical evaluation, using simulations in a clinical setting and real-world data from the child welfare domain, indicates that the proposed approach successfully narrows the construct gap.
arXiv Detail & Related papers (2021-01-24T05:40:29Z) - FrugalML: How to Use ML Prediction APIs More Accurately and Cheaply [36.94826820536239]
We propose FrugalML, a principled framework that jointly learns the strength and weakness of each API on different data.
Our theoretical analysis shows that natural sparsity in the formulation can be leveraged to make FrugalML efficient.
Across various tasks, FrugalML can achieve up to 90% cost reduction while matching the accuracy of the best single API, or up to 5% better accuracy while matching the best API's cost.
arXiv Detail & Related papers (2020-06-12T23:43:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.