Beimingwu: A Learnware Dock System
- URL: http://arxiv.org/abs/2401.14427v1
- Date: Wed, 24 Jan 2024 09:27:51 GMT
- Title: Beimingwu: A Learnware Dock System
- Authors: Zhi-Hao Tan, Jian-Dong Liu, Xiao-Dong Bi, Peng Tan, Qin-Cheng Zheng,
Hai-Tian Liu, Yi Xie, Xiao-Chuan Zou, Yang Yu, Zhi-Hua Zhou
- Abstract summary: This paper describes Beimingwu, the first open-source learnware dock system providing foundational support for future research of learnware paradigm.
The system significantly streamlines the model development for new user tasks, thanks to its integrated architecture and engine design.
Notably, this is possible even for users with limited data and minimal expertise in machine learning, without compromising the raw data's security.
- Score: 42.54363998206648
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The learnware paradigm proposed by Zhou [2016] aims to enable users to reuse
numerous existing well-trained models instead of building machine learning
models from scratch, with the hope of solving new user tasks even beyond
models' original purposes. In this paradigm, developers worldwide can submit
their high-performing models spontaneously to the learnware dock system
(formerly known as learnware market) without revealing their training data.
Once the dock system accepts the model, it assigns a specification and
accommodates the model. This specification allows the model to be adequately
identified and assembled to reuse according to future users' needs, even if
they have no prior knowledge of the model. This paradigm greatly differs from
the current big model direction and it is expected that a learnware dock system
housing millions or more high-performing models could offer excellent
capabilities for both planned tasks where big models are applicable; and
unplanned, specialized, data-sensitive scenarios where big models are not
present or applicable.
This paper describes Beimingwu, the first open-source learnware dock system
providing foundational support for future research of learnware paradigm.The
system significantly streamlines the model development for new user tasks,
thanks to its integrated architecture and engine design, extensive engineering
implementations and optimizations, and the integration of various algorithms
for learnware identification and reuse. Notably, this is possible even for
users with limited data and minimal expertise in machine learning, without
compromising the raw data's security. Beimingwu supports the entire process of
learnware paradigm. The system lays the foundation for future research in
learnware-related algorithms and systems, and prepares the ground for hosting a
vast array of learnwares and establishing a learnware ecosystem.
Related papers
- Reusable MLOps: Reusable Deployment, Reusable Infrastructure and
Hot-Swappable Machine Learning models and services [0.0]
We introduce a new sustainable concept in the field of AI/ML operations - called Reusable MLOps.
We reuse the existing deployment and infrastructure to serve new models by hot-swapping them without tearing down the infrastructure or the microservice.
arXiv Detail & Related papers (2024-02-19T23:40:46Z) - ZhiJian: A Unifying and Rapidly Deployable Toolbox for Pre-trained Model
Reuse [59.500060790983994]
This paper introduces ZhiJian, a comprehensive and user-friendly toolbox for model reuse, utilizing the PyTorch backend.
ZhiJian presents a novel paradigm that unifies diverse perspectives on model reuse, encompassing target architecture construction with PTM, tuning target model with PTM, and PTM-based inference.
arXiv Detail & Related papers (2023-08-17T19:12:13Z) - Model Ratatouille: Recycling Diverse Models for Out-of-Distribution
Generalization [99.6826401545377]
Foundation models are redefining how AI systems are built. Practitioners now follow a standard procedure to build their machine learning solutions.
We propose model ratatouille, a new strategy to recycle the multiple fine-tunings of the same foundation model on diverse auxiliary tasks.
arXiv Detail & Related papers (2022-12-20T17:21:46Z) - On-device Training: A First Overview on Existing Systems [6.551096686706628]
Efforts have been made to deploy some models on resource-constrained devices as well.
This work targets to summarize and analyze state-of-the-art systems research that allows such on-device model training capabilities.
arXiv Detail & Related papers (2022-12-01T19:22:29Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - Learnware: Small Models Do Big [69.88234743773113]
The prevailing big model paradigm, which has achieved impressive results in natural language processing and computer vision applications, has not yet addressed those issues, whereas becoming a serious source of carbon emissions.
This article offers an overview of the learnware paradigm, which attempts to enable users not need to build machine learning models from scratch, with the hope of reusing small models to do things even beyond their original purposes.
arXiv Detail & Related papers (2022-10-07T15:55:52Z) - Retrieval-Enhanced Machine Learning [110.5237983180089]
We describe a generic retrieval-enhanced machine learning framework, which includes a number of existing models as special cases.
REML challenges information retrieval conventions, presenting opportunities for novel advances in core areas, including optimization.
REML research agenda lays a foundation for a new style of information access research and paves a path towards advancing machine learning and artificial intelligence.
arXiv Detail & Related papers (2022-05-02T21:42:45Z) - A Model-Driven Engineering Approach to Machine Learning and Software
Modeling [0.5156484100374059]
Models are used in both the Software Engineering (SE) and the Artificial Intelligence (AI) communities.
The main focus is on the Internet of Things (IoT) and smart Cyber-Physical Systems (CPS) use cases, where both ML and model-driven SE play a key role.
arXiv Detail & Related papers (2021-07-06T15:50:50Z) - Green Lighting ML: Confidentiality, Integrity, and Availability of
Machine Learning Systems in Deployment [4.2317391919680425]
In production machine learning, there is generally a hand-off from those who build a model to those who deploy a model.
In this hand-off, the engineers responsible for model deployment are often not privy to the details of the model.
In order to help alleviate this issue, automated systems for validating privacy and security of models need to be developed.
arXiv Detail & Related papers (2020-07-09T10:38:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.