Studying the Practices of Deploying Machine Learning Projects on Docker
- URL: http://arxiv.org/abs/2206.00699v1
- Date: Wed, 1 Jun 2022 18:13:30 GMT
- Title: Studying the Practices of Deploying Machine Learning Projects on Docker
- Authors: Moses Openja, Forough Majidi, Foutse Khomh, Bhagya Chembakottu, Heng
Li
- Abstract summary: Docker is a containerization service that allows for convenient deployment of websites, databases, applications' APIs, and machine learning (ML) models with a few lines of code.
We conducted an exploratory study to understand how Docker is being used to deploy ML-based projects.
- Score: 9.979005459305117
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Docker is a containerization service that allows for convenient deployment of
websites, databases, applications' APIs, and machine learning (ML) models with
a few lines of code. Studies have recently explored the use of Docker for
deploying general software projects with no specific focus on how Docker is
used to deploy ML-based projects.
In this study, we conducted an exploratory study to understand how Docker is
being used to deploy ML-based projects. As the initial step, we examined the
categories of ML-based projects that use Docker. We then examined why and how
these projects use Docker, and the characteristics of the resulting Docker
images. Our results indicate that six categories of ML-based projects use
Docker for deployment, including ML Applications, MLOps/ AIOps, Toolkits, DL
Frameworks, Models, and Documentation. We derived the taxonomy of 21 major
categories representing the purposes of using Docker, including those specific
to models such as model management tasks (e.g., testing, training). We then
showed that ML engineers use Docker images mostly to help with the platform
portability, such as transferring the software across the operating systems,
runtimes such as GPU, and language constraints. However, we also found that
more resources may be required to run the Docker images for building ML-based
software projects due to the large number of files contained in the image
layers with deeply nested directories. We hope to shed light on the emerging
practices of deploying ML software projects using containers and highlight
aspects that should be improved.
Related papers
- Toward Automated Test Generation for Dockerfiles Based on Analysis of Docker Image Layers [1.1879716317856948]
The process for building a Docker image is defined in a text file called a Dockerfile.
A Dockerfile can be considered as a kind of source code that contains instructions on how to build a Docker image.
We propose an automated test generation method for Dockerfiles based on processing results rather than processing steps.
arXiv Detail & Related papers (2025-04-25T08:02:46Z) - Design and Implementation of Flutter based Multi-platform Docker Controller App [1.1443262816483672]
This paper focuses on developing a Flutter application for controlling Docker resources remotely.
The application uses the SSH protocol to establish a secure connection with the server and execute the commands.
An alternative approach is also explored, which involves connecting the application with the Docker engine using HTTP.
arXiv Detail & Related papers (2025-02-17T11:48:02Z) - Refactoring for Dockerfile Quality: A Dive into Developer Practices and Automation Potential [0.0]
This paper explores the utility and practicality of automating Dockerfile using 600files from 358 open-source projects.
Our approach leads to an average reduction of 32% in image size and a 6% decrease in build duration, with improvements in understandability and maintainability observed in 77% and 91% of cases.
arXiv Detail & Related papers (2025-01-23T23:10:47Z) - Task Preference Optimization: Improving Multimodal Large Language Models with Vision Task Alignment [58.94611347128066]
Task Preference Optimization (TPO) is a novel method that utilizes differentiable task preferences derived from typical fine-grained visual tasks.
By leveraging rich visual labels during training, TPO significantly enhances the MLLM's multimodal capabilities and task-specific performance.
Our instantiation of this approach with VideoChat and LLaVA demonstrates an overall 14.6% improvement in multimodal performance compared to baseline models.
arXiv Detail & Related papers (2024-12-26T18:56:05Z) - Teaching VLMs to Localize Specific Objects from In-context Examples [56.797110842152]
Vision-Language Models (VLMs) have shown remarkable capabilities across diverse visual tasks.
Current VLMs lack a fundamental cognitive ability: learning to localize objects in a scene by taking into account the context.
This work is the first to explore and benchmark personalized few-shot localization for VLMs.
arXiv Detail & Related papers (2024-11-20T13:34:22Z) - On the Cost of Model-Serving Frameworks: An Experimental Evaluation [2.6232657671486983]
Serving strategies are crucial for deploying and managing models in production environments effectively.
These strategies ensure that models are available, scalable, reliable, and performant for real-world applications.
We show that DL-specific frameworks (TensorFlow Serving and TorchServe) display significantly lower latencies than the three general-purpose ML frameworks.
arXiv Detail & Related papers (2024-11-15T16:36:21Z) - MOWA: Multiple-in-One Image Warping Model [65.73060159073644]
We propose a Multiple-in-One image warping model (named MOWA) in this work.
We mitigate the difficulty of multi-task learning by disentangling the motion estimation at both the region level and pixel level.
To our knowledge, this is the first work that solves multiple practical warping tasks in one single model.
arXiv Detail & Related papers (2024-04-16T16:50:35Z) - General Object Foundation Model for Images and Videos at Scale [99.2806103051613]
We present GLEE, an object-level foundation model for locating and identifying objects in images and videos.
GLEE accomplishes detection, segmentation, tracking, grounding, and identification of arbitrary objects in the open world scenario.
We employ an image encoder, text encoder, and visual prompter to handle multi-modal inputs, enabling to simultaneously solve various object-centric downstream tasks.
arXiv Detail & Related papers (2023-12-14T17:26:00Z) - ML-Bench: Evaluating Large Language Models and Agents for Machine Learning Tasks on Repository-Level Code [76.84199699772903]
ML-Bench is a benchmark rooted in real-world programming applications that leverage existing code repositories to perform tasks.
To evaluate both Large Language Models (LLMs) and AI agents, two setups are employed: ML-LLM-Bench for assessing LLMs' text-to-code conversion within a predefined deployment environment, and ML-Agent-Bench for testing autonomous agents in an end-to-end task execution within a Linux sandbox environment.
arXiv Detail & Related papers (2023-11-16T12:03:21Z) - CRAFT: Customizing LLMs by Creating and Retrieving from Specialized
Toolsets [75.64181719386497]
We present CRAFT, a tool creation and retrieval framework for large language models (LLMs)
It creates toolsets specifically curated for the tasks and equips LLMs with a component that retrieves tools from these sets to enhance their capability to solve complex tasks.
Our method is designed to be flexible and offers a plug-and-play approach to adapt off-the-shelf LLMs to unseen domains and modalities, without any finetuning.
arXiv Detail & Related papers (2023-09-29T17:40:26Z) - MLOps: A Step Forward to Enterprise Machine Learning [0.0]
This research presents a detailed review of MLOps, its benefits, difficulties, evolutions, and important underlying technologies.
The MLOps workflow is explained in detail along with the various tools necessary for both model and data exploration and deployment.
This article also puts light on the end-to-end production of ML projects using various maturity levels of automated pipelines.
arXiv Detail & Related papers (2023-05-27T20:44:14Z) - Patterns of Multi-Container Composition for Service Orchestration with Docker Compose [5.1861106408299635]
This work curates a dataset of successful projects that employ Docker Compose as an orchestration tool to run multiple service containers.
The collection of data and analysis enables the identification and naming of repeating multi-container composition patterns.
These patterns highlight how software systems are orchestrated in the real-world and can give examples to anybody wishing to compose their own service orchestrations.
arXiv Detail & Related papers (2023-05-18T20:32:58Z) - VisionLLM: Large Language Model is also an Open-Ended Decoder for
Vision-Centric Tasks [81.32968995346775]
VisionLLM is a framework for vision-centric tasks that can be flexibly defined and managed using language instructions.
Our model can achieve over 60% mAP on COCO, on par with detection-specific models.
arXiv Detail & Related papers (2023-05-18T17:59:42Z) - A Preliminary Investigation of MLOps Practices in GitHub [10.190501703364234]
Machine learning applications have led to an increasing interest in MLOps.
We present an initial investigation of the MLOps practices implemented in a set of ML-enabled systems retrieved from GitHub.
arXiv Detail & Related papers (2022-09-23T07:29:56Z) - MLModelCI: An Automatic Cloud Platform for Efficient MLaaS [15.029094196394862]
We release the platform as an open-source project on GitHub under Apache 2.0 license.
Our system bridges the gap between current ML training and serving systems and thus free developers from manual and tedious work often associated with service deployment.
arXiv Detail & Related papers (2020-06-09T07:48:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.