ModelPS: An Interactive and Collaborative Platform for Editing
Pre-trained Models at Scale
- URL: http://arxiv.org/abs/2105.08275v1
- Date: Tue, 18 May 2021 04:51:56 GMT
- Title: ModelPS: An Interactive and Collaborative Platform for Editing
Pre-trained Models at Scale
- Authors: Yuanming Li, Huaizheng Zhang, Shanshan Jiang, Fan Yang, Yonggang Wen
and Yong Luo
- Abstract summary: We propose and develop a low-code solution, ModelPS, to enable collaborative DNN model editing and intelligent model serving.
The ModelPS solution embodies two transformative features: 1) a user-friendly web interface for a developer team to share and edit DNN models pictorially, in a low-code fashion, and 2) a model genie engine in the backend to aid developers in customizing model editing configurations for given deployment requirements or constraints.
- Score: 30.333660470820604
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: AI engineering has emerged as a crucial discipline to democratize deep neural
network (DNN) models among software developers with a diverse background. In
particular, altering these DNN models in the deployment stage posits a
tremendous challenge. In this research, we propose and develop a low-code
solution, ModelPS (an acronym for "Model Photoshop"), to enable and empower
collaborative DNN model editing and intelligent model serving. The ModelPS
solution embodies two transformative features: 1) a user-friendly web interface
for a developer team to share and edit DNN models pictorially, in a low-code
fashion, and 2) a model genie engine in the backend to aid developers in
customizing model editing configurations for given deployment requirements or
constraints. Our case studies with a wide range of deep learning (DL) models
show that the system can tremendously reduce both development and communication
overheads with improved productivity. The code has been released as an
open-source package at GitHub.
Related papers
- Qwen2-VL: Enhancing Vision-Language Model's Perception of the World at Any Resolution [82.38677987249348]
We present the Qwen2-VL Series, which redefines the conventional predetermined-resolution approach in visual processing.
Qwen2-VL introduces the Naive Dynamic Resolution mechanism, which enables the model to dynamically process images of varying resolutions into different numbers of visual tokens.
The model also integrates Multimodal Rotary Position Embedding (M-RoPE), facilitating the effective fusion of positional information across text, images, and videos.
arXiv Detail & Related papers (2024-09-18T17:59:32Z) - Apple Intelligence Foundation Language Models [109.60033785567484]
This report describes the model architecture, the data used to train the model, the training process, and the evaluation results.
We highlight our focus on Responsible AI and how the principles are applied throughout the model development.
arXiv Detail & Related papers (2024-07-29T18:38:49Z) - NVC-1B: A Large Neural Video Coding Model [36.66792728623893]
We design the first neural video coding model with more than 1 billion parameters -- NVC-1B.
Experimental results show that our proposed large model achieves a significant video compression performance improvement.
We anticipate large models may bring up the video coding technologies to the next level.
arXiv Detail & Related papers (2024-07-28T05:12:22Z) - Model Callers for Transforming Predictive and Generative AI Applications [2.7195102129095003]
We introduce a novel software abstraction termed "model caller"
Model callers act as an intermediary for AI and ML model calling.
We have released a prototype Python library for model callers, accessible for installation via pip or for download from GitHub.
arXiv Detail & Related papers (2024-04-17T12:21:06Z) - Towards Automatic Support of Software Model Evolution with Large
Language~Models [6.872484164111954]
We propose an approach that utilizes large language models for model completion and discovering editing patterns in model histories of software systems.
We have found that large language models are indeed a promising technology for supporting software model evolution.
arXiv Detail & Related papers (2023-12-19T18:38:01Z) - FlexModel: A Framework for Interpretability of Distributed Large
Language Models [0.0]
We present FlexModel, a software package providing a streamlined interface for engaging with models distributed across multi- GPU and multi-node configurations.
The library is compatible with existing model distribution libraries and encapsulates PyTorch models.
It exposes user-registerable HookFunctions to facilitate straightforward interaction with distributed model internals.
arXiv Detail & Related papers (2023-12-05T21:19:33Z) - UnIVAL: Unified Model for Image, Video, Audio and Language Tasks [105.77733287326308]
UnIVAL model goes beyond two modalities and unifies text, images, video, and audio into a single model.
Our model is efficiently pretrained on many tasks, based on task balancing and multimodal curriculum learning.
Thanks to the unified model, we propose a novel study on multimodal model merging via weight generalization.
arXiv Detail & Related papers (2023-07-30T09:48:36Z) - Speculative Decoding with Big Little Decoder [108.95187338417541]
Big Little Decoder (BiLD) is a framework that can improve inference efficiency and latency for a wide range of text generation applications.
On an NVIDIA T4 GPU, our framework achieves a speedup of up to 2.12x speedup with minimal generation quality degradation.
Our framework is fully plug-and-play and can be applied without any modifications in the training process or model architecture.
arXiv Detail & Related papers (2023-02-15T18:55:29Z) - Model LEGO: Creating Models Like Disassembling and Assembling Building Blocks [53.09649785009528]
In this paper, we explore a paradigm that does not require training to obtain new models.
Similar to the birth of CNN inspired by receptive fields in the biological visual system, we propose Model Disassembling and Assembling.
For model assembling, we present the alignment padding strategy and parameter scaling strategy to construct a new model tailored for a specific task.
arXiv Detail & Related papers (2022-03-25T05:27:28Z) - JMSNAS: Joint Model Split and Neural Architecture Search for Learning
over Mobile Edge Networks [23.230079759174902]
Joint model split and neural architecture search (JMSNAS) framework is proposed to automatically generate and deploy a DNN model over a mobile edge network.
Considering both the computing and communication resource constraints, a computational graph search problem is formulated.
Experiment results confirm the superiority of the proposed framework over the state-of-the-art split machine learning design methods.
arXiv Detail & Related papers (2021-11-16T03:10:23Z) - Data-Driven and SE-assisted AI Model Signal-Awareness Enhancement and
Introspection [61.571331422347875]
We propose a data-driven approach to enhance models' signal-awareness.
We combine the SE concept of code complexity with the AI technique of curriculum learning.
We achieve up to 4.8x improvement in model signal awareness.
arXiv Detail & Related papers (2021-11-10T17:58:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.