ApproxIFER: A Model-Agnostic Approach to Resilient and Robust Prediction
Serving Systems
- URL: http://arxiv.org/abs/2109.09868v1
- Date: Mon, 20 Sep 2021 22:29:57 GMT
- Title: ApproxIFER: A Model-Agnostic Approach to Resilient and Robust Prediction
Serving Systems
- Authors: Mahdi Soleymani, Ramy E. Ali, Hessam Mahdavifar, A. Salman Avestimehr
- Abstract summary: We propose ApproxIFER to handle a general number of stragglers and scales significantly better with the number of queries.
Our experiments on a large number of datasets and model architectures also show significant accuracy improvement by up to 58% over the parity model approaches.
- Score: 24.75623641870649
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to the surge of cloud-assisted AI services, the problem of designing
resilient prediction serving systems that can effectively cope with
stragglers/failures and minimize response delays has attracted much interest.
The common approach for tackling this problem is replication which assigns the
same prediction task to multiple workers. This approach, however, is very
inefficient and incurs significant resource overheads. Hence, a learning-based
approach known as parity model (ParM) has been recently proposed which learns
models that can generate parities for a group of predictions in order to
reconstruct the predictions of the slow/failed workers. While this
learning-based approach is more resource-efficient than replication, it is
tailored to the specific model hosted by the cloud and is particularly suitable
for a small number of queries (typically less than four) and tolerating very
few (mostly one) number of stragglers. Moreover, ParM does not handle Byzantine
adversarial workers. We propose a different approach, named Approximate Coded
Inference (ApproxIFER), that does not require training of any parity models,
hence it is agnostic to the model hosted by the cloud and can be readily
applied to different data domains and model architectures. Compared with
earlier works, ApproxIFER can handle a general number of stragglers and scales
significantly better with the number of queries. Furthermore, ApproxIFER is
robust against Byzantine workers. Our extensive experiments on a large number
of datasets and model architectures also show significant accuracy improvement
by up to 58% over the parity model approaches.
Related papers
- A Federated Data Fusion-Based Prognostic Model for Applications with Multi-Stream Incomplete Signals [1.2277343096128712]
This article proposes a federated prognostic model that allows multiple users to jointly construct a failure time prediction model.
Numerical studies indicate that the performance of the proposed model is the same as that of classic non-federated prognostic models.
arXiv Detail & Related papers (2023-11-13T17:08:34Z) - Structured Radial Basis Function Network: Modelling Diversity for
Multiple Hypotheses Prediction [51.82628081279621]
Multi-modal regression is important in forecasting nonstationary processes or with a complex mixture of distributions.
A Structured Radial Basis Function Network is presented as an ensemble of multiple hypotheses predictors for regression problems.
It is proved that this structured model can efficiently interpolate this tessellation and approximate the multiple hypotheses target distribution.
arXiv Detail & Related papers (2023-09-02T01:27:53Z) - Window-Based Early-Exit Cascades for Uncertainty Estimation: When Deep
Ensembles are More Efficient than Single Models [5.0401589279256065]
We show that ensembles can be more computationally efficient (at inference) than scaling single models within an architecture family.
In this work, we investigate extending these efficiency gains to tasks related to uncertainty estimation.
Experiments on ImageNet-scale data across a number of network architectures and uncertainty tasks show that the proposed window-based early-exit approach is able to achieve a superior uncertainty-computation trade-off.
arXiv Detail & Related papers (2023-03-14T15:57:54Z) - Scalable Estimation for Structured Additive Distributional Regression [0.0]
We propose a novel backfitting algorithm, which is based on the ideas of gradient descent and can deal virtually with any amount of data on a conventional laptop.
Performance is evaluated using an extensive simulation study and an exceptionally challenging and unique example of lightning count prediction over Austria.
arXiv Detail & Related papers (2023-01-13T14:59:42Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - DualCF: Efficient Model Extraction Attack from Counterfactual
Explanations [57.46134660974256]
Cloud service providers have launched Machine-Learning-as-a-Service platforms to allow users to access large-scale cloudbased models via APIs.
Such extra information inevitably causes the cloud models to be more vulnerable to extraction attacks.
We propose a novel simple yet efficient querying strategy to greatly enhance the querying efficiency to steal a classification model.
arXiv Detail & Related papers (2022-05-13T08:24:43Z) - An Expectation-Maximization Perspective on Federated Learning [75.67515842938299]
Federated learning describes the distributed training of models across multiple clients while keeping the data private on-device.
In this work, we view the server-orchestrated federated learning process as a hierarchical latent variable model where the server provides the parameters of a prior distribution over the client-specific model parameters.
We show that with simple Gaussian priors and a hard version of the well known Expectation-Maximization (EM) algorithm, learning in such a model corresponds to FedAvg, the most popular algorithm for the federated learning setting.
arXiv Detail & Related papers (2021-11-19T12:58:59Z) - Kronecker Factorization for Preventing Catastrophic Forgetting in
Large-scale Medical Entity Linking [7.723047334864811]
In the medical domain, sequential training on tasks may sometimes be the only way to train models.
catastrophic forgetting, i.e., a substantial drop in accuracy on prior tasks when a model is updated for a new task.
We show the effectiveness of this technique on the important and illustrative task of medical entity linking across three datasets.
arXiv Detail & Related papers (2021-11-11T01:51:01Z) - BODAME: Bilevel Optimization for Defense Against Model Extraction [10.877450596327407]
We consider an adversarial setting to prevent model extraction under the assumption that will make best guess on the service provider's attacker.
We formulate a surrogate model using the predictions of the true model.
We give a tractable transformation and an algorithm for more complicated models that are learned by using gradient descent-based algorithms.
arXiv Detail & Related papers (2021-03-11T17:08:31Z) - Ambiguity in Sequential Data: Predicting Uncertain Futures with
Recurrent Models [110.82452096672182]
We propose an extension of the Multiple Hypothesis Prediction (MHP) model to handle ambiguous predictions with sequential data.
We also introduce a novel metric for ambiguous problems, which is better suited to account for uncertainties.
arXiv Detail & Related papers (2020-03-10T09:15:42Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.