Parallelizing Machine Learning as a Service for the End-User
- URL: http://arxiv.org/abs/2005.14080v2
- Date: Fri, 29 May 2020 09:13:36 GMT
- Title: Parallelizing Machine Learning as a Service for the End-User
- Authors: Daniela Loreti and Marco Lippi and Paolo Torroni
- Abstract summary: We present a distributed architecture that could be exploited to parallelize a typical ML system pipeline.
We propose a case study consisting of a text mining service and discuss how the method can be generalized to many similar applications.
- Score: 14.389966909395058
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As ML applications are becoming ever more pervasive, fully-trained systems
are made increasingly available to a wide public, allowing end-users to submit
queries with their own data, and to efficiently retrieve results. With
increasingly sophisticated such services, a new challenge is how to scale up to
evergrowing user bases. In this paper, we present a distributed architecture
that could be exploited to parallelize a typical ML system pipeline. We propose
a case study consisting of a text mining service and discuss how the method can
be generalized to many similar applications. We demonstrate the significance of
the computational gain boosted by the distributed architecture by way of an
extensive experimental evaluation.
Related papers
- Dynamic Few-Shot Learning for Knowledge Graph Question Answering [3.116231004560997]
Large language models present opportunities for innovative Question Answering over Knowledge Graphs (KGQA)
To bridge this gap, solutions have been proposed that rely on fine-tuning or ad-hoc architectures, achieving good results but limited out-of-domain distribution generalization.
In this study, we introduce a novel approach called Dynamic Few-Shot Learning (DFL)
DFL integrates the efficiency of in-context learning and semantic similarity and provides a generally applicable solution for KGQA with state-of-the-art performance.
arXiv Detail & Related papers (2024-07-01T15:59:17Z) - Expert Router: Orchestrating Efficient Language Model Inference through Prompt Classification [0.4726677580049183]
Large Language Models (LLMs) have experienced widespread adoption across scientific and industrial domains.
We introduce Expert Router, a system designed to orchestrate multiple expert models efficiently.
Expert Router is a parallel inference system with a central routing gateway that distributes incoming requests.
arXiv Detail & Related papers (2024-04-22T16:33:42Z) - Serving Deep Learning Model in Relational Databases [72.72372281808694]
Serving deep learning (DL) models on relational data has become a critical requirement across diverse commercial and scientific domains.
We highlight three pivotal paradigms: The state-of-the-artDL-Centric architecture offloadsDL computations to dedicated DL frameworks.
The potential UDF-Centric architecture encapsulates one or more tensor computations into User Defined Functions (UDFs) within the database system.
The potentialRelation-Centric architecture aims to represent a large-scale tensor computation through operators.
arXiv Detail & Related papers (2023-10-07T06:01:35Z) - Architect, Regularize and Replay (ARR): a Flexible Hybrid Approach for
Continual Learning [13.492896179777835]
"Architect, Regularize and Replay" (ARR) is a hybrid generalization of the renowned AR1 algorithm and its variants.
It can achieve state-of-the-art results in classic scenarios (e.g. class-incremental learning) but also generalize to arbitrary data streams generated from real-world datasets such as CIFAR-100, CORe50 and ImageNet-1000.
arXiv Detail & Related papers (2023-01-06T11:22:59Z) - Distributed intelligence on the Edge-to-Cloud Continuum: A systematic
literature review [62.997667081978825]
This review aims at providing a comprehensive vision of the main state-of-the-art libraries and frameworks for machine learning and data analytics available today.
The main simulation, emulation, deployment systems, and testbeds for experimental research on the Edge-to-Cloud Continuum available today are also surveyed.
arXiv Detail & Related papers (2022-04-29T08:06:05Z) - SOLIS -- The MLOps journey from data acquisition to actionable insights [62.997667081978825]
In this paper we present a unified deployment pipeline and freedom-to-operate approach that supports all requirements while using basic cross-platform tensor framework and script language engines.
This approach however does not supply the needed procedures and pipelines for the actual deployment of machine learning capabilities in real production grade systems.
arXiv Detail & Related papers (2021-12-22T14:45:37Z) - Deployment Optimization for Shared e-Mobility Systems with Multi-agent
Deep Neural Search [15.657420177295624]
Shared e-mobility services have been widely tested and piloted in cities across the globe.
This paper studies how to deploy and manage their infrastructure across space and time, so that the services are ubiquitous to the users while in sustainable profitability.
We tackle this by designing a high-fidelity simulation environment, which abstracts the key operation details of the shared e-mobility systems at fine-granularity.
arXiv Detail & Related papers (2021-11-03T11:37:11Z) - Edge-assisted Democratized Learning Towards Federated Analytics [67.44078999945722]
We show the hierarchical learning structure of the proposed edge-assisted democratized learning mechanism, namely Edge-DemLearn.
We also validate Edge-DemLearn as a flexible model training mechanism to build a distributed control and aggregation methodology in regions.
arXiv Detail & Related papers (2020-12-01T11:46:03Z) - A Survey on Large-scale Machine Learning [67.6997613600942]
Machine learning can provide deep insights into data, allowing machines to make high-quality predictions.
Most sophisticated machine learning approaches suffer from huge time costs when operating on large-scale data.
Large-scale Machine Learning aims to learn patterns from big data with comparable performance efficiently.
arXiv Detail & Related papers (2020-08-10T06:07:52Z) - A Privacy-Preserving Distributed Architecture for
Deep-Learning-as-a-Service [68.84245063902908]
This paper introduces a novel distributed architecture for deep-learning-as-a-service.
It is able to preserve the user sensitive data while providing Cloud-based machine and deep learning services.
arXiv Detail & Related papers (2020-03-30T15:12:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.