Design Patterns for Machine Learning Based Systems with
Human-in-the-Loop
- URL: http://arxiv.org/abs/2312.00582v1
- Date: Fri, 1 Dec 2023 13:46:38 GMT
- Title: Design Patterns for Machine Learning Based Systems with
Human-in-the-Loop
- Authors: Jakob Smedegaard Andersen and Walid Maalej
- Abstract summary: Humans involvement in machine learning is a promising and powerful paradigm to overcome the limitations of pure automated predictions.
We compile a catalog of design patterns to guide developers select and implement suitable human-in-the-loop (HiL) solutions.
- Score: 13.720835527532733
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The development and deployment of systems using supervised machine learning
(ML) remain challenging: mainly due to the limited reliability of prediction
models and the lack of knowledge on how to effectively integrate human
intelligence into automated decision-making. Humans involvement in the ML
process is a promising and powerful paradigm to overcome the limitations of
pure automated predictions and improve the applicability of ML in practice. We
compile a catalog of design patterns to guide developers select and implement
suitable human-in-the-loop (HiL) solutions. Our catalog takes into
consideration key requirements as the cost of human involvement and model
retraining. It includes four training patterns, four deployment patterns, and
two orthogonal cooperation patterns.
Related papers
- To Ask or Not to Ask: Learning to Require Human Feedback [16.806124909744877]
We propose a new framework that handles both when and how to incorporate expert input in an Machine Learning model.<n>LtA is based on a two-part architecture: a standard ML model and an enriched model trained with additional expert human feedback.<n>We provide two practical implementations of LtA: a sequential approach, which trains the models in stages, and a joint approach, which optimises them simultaneously.
arXiv Detail & Related papers (2025-10-09T15:00:06Z) - MoSE: Skill-by-Skill Mixture-of-Expert Learning for Autonomous Driving [14.042949333988785]
We propose a skill-oriented MoE, called MoSE, which mimics human drivers' learning process, skill-by-skill and step-by-step.<n>We build a hierarchical skill dataset and pretrain the router to encourage the model to think step-by-step.<n>With less than 3B sparsely activated parameters, our model outperforms several 8B+ parameters on CODA AD corner case reasoning task.
arXiv Detail & Related papers (2025-07-10T14:48:08Z) - Continual Learning for Generative AI: From LLMs to MLLMs and Beyond [56.29231194002407]
We present a comprehensive survey of continual learning methods for mainstream generative AI models.<n>We categorize these approaches into three paradigms: architecture-based, regularization-based, and replay-based.<n>We analyze continual learning setups for different generative models, including training objectives, benchmarks, and core backbones.
arXiv Detail & Related papers (2025-06-16T02:27:25Z) - Approach to Finding a Robust Deep Learning Model [0.28675177318965045]
The rapid development of machine learning (ML) and artificial intelligence (AI) applications requires the training of large numbers of models.<n>We propose a novel approach for determining model robustness using a proposed model selection algorithm designed as a meta-algorithm.<n>Within this framework, we address the influence of training sample size, model weight, and inductive bias on the robustness of deep learning models.
arXiv Detail & Related papers (2025-05-22T20:05:20Z) - Human-In-The-Loop Machine Learning for Safe and Ethical Autonomous Vehicles: Principles, Challenges, and Opportunities [33.853994070508485]
We focus on Curriculum Learning (CL), Human-In-The-Loop Reinforcement Learning (HITL-RL), Active Learning (AL), and ethical principles.
In CL, human experts systematically train ML models by starting with simple tasks and gradually progressing to more difficult ones.
HITL-RL significantly enhances the RL process by incorporating human input through techniques like reward shaping, action injection, and interactive learning.
AL streamlines the annotation process by targeting specific instances that need to be labeled with human oversight.
arXiv Detail & Related papers (2024-08-22T17:02:29Z) - PersLLM: A Personified Training Approach for Large Language Models [66.16513246245401]
We propose PersLLM, a framework for better data construction and model tuning.<n>For insufficient data usage, we incorporate strategies such as Chain-of-Thought prompting and anti-induction.<n>For rigid behavior patterns, we design the tuning process and introduce automated DPO to enhance the specificity and dynamism of the models' personalities.
arXiv Detail & Related papers (2024-07-17T08:13:22Z) - Predicting and Understanding Human Action Decisions: Insights from Large Language Models and Cognitive Instance-Based Learning [0.0]
Large Language Models (LLMs) have demonstrated their capabilities across various tasks.
This paper exploits the reasoning and generative capabilities of the LLMs to predict human behavior in two sequential decision-making tasks.
We compare the performance of LLMs with a cognitive instance-based learning model, which imitates human experiential decision-making.
arXiv Detail & Related papers (2024-07-12T14:13:06Z) - Machine Learning Insides OptVerse AI Solver: Design Principles and
Applications [74.67495900436728]
We present a comprehensive study on the integration of machine learning (ML) techniques into Huawei Cloud's OptVerse AI solver.
We showcase our methods for generating complex SAT and MILP instances utilizing generative models that mirror multifaceted structures of real-world problem.
We detail the incorporation of state-of-the-art parameter tuning algorithms which markedly elevate solver performance.
arXiv Detail & Related papers (2024-01-11T15:02:15Z) - TrainerAgent: Customizable and Efficient Model Training through
LLM-Powered Multi-Agent System [14.019244136838017]
TrainerAgent is a multi-agent framework including Task, Data, Model and Server agents.
These agents analyze user-defined tasks, input data, and requirements (e.g., accuracy, speed), optimizing them from both data and model perspectives to obtain satisfactory models, and finally deploy these models as online service.
This research presents a significant advancement in achieving desired models with increased efficiency and quality as compared to traditional model development.
arXiv Detail & Related papers (2023-11-11T17:39:24Z) - Self-Destructing Models: Increasing the Costs of Harmful Dual Uses of
Foundation Models [103.71308117592963]
We present an algorithm for training self-destructing models leveraging techniques from meta-learning and adversarial learning.
In a small-scale experiment, we show MLAC can largely prevent a BERT-style model from being re-purposed to perform gender identification.
arXiv Detail & Related papers (2022-11-27T21:43:45Z) - Tiny Robot Learning: Challenges and Directions for Machine Learning in
Resource-Constrained Robots [57.27442333662654]
Machine learning (ML) has become a pervasive tool across computing systems.
Tiny robot learning is the deployment of ML on resource-constrained low-cost autonomous robots.
Tiny robot learning is subject to challenges from size, weight, area, and power (SWAP) constraints.
This paper gives a brief survey of the tiny robot learning space, elaborates on key challenges, and proposes promising opportunities for future work in ML system design.
arXiv Detail & Related papers (2022-05-11T19:36:15Z) - Enabling Automated Machine Learning for Model-Driven AI Engineering [60.09869520679979]
We propose a novel approach to enable Model-Driven Software Engineering and Model-Driven AI Engineering.
In particular, we support Automated ML, thus assisting software engineers without deep AI knowledge in developing AI-intensive systems.
arXiv Detail & Related papers (2022-03-06T10:12:56Z) - Self-directed Machine Learning [86.3709575146414]
In education science, self-directed learning has been shown to be more effective than passive teacher-guided learning.
We introduce the principal concept of Self-directed Machine Learning (SDML) and propose a framework for SDML.
Our proposed SDML process benefits from self task selection, self data selection, self model selection, self optimization strategy selection and self evaluation metric selection.
arXiv Detail & Related papers (2022-01-04T18:32:06Z) - Panoramic Learning with A Standardized Machine Learning Formalism [116.34627789412102]
This paper presents a standardized equation of the learning objective, that offers a unifying understanding of diverse ML algorithms.
It also provides guidance for mechanic design of new ML solutions, and serves as a promising vehicle towards panoramic learning with all experiences.
arXiv Detail & Related papers (2021-08-17T17:44:38Z) - Towards Model-informed Precision Dosing with Expert-in-the-loop Machine
Learning [0.0]
We consider a ML framework that may accelerate model learning and improve its interpretability by incorporating human experts into the model learning loop.
We propose a novel human-in-the-loop ML framework aimed at dealing with learning problems that the cost of data annotation is high.
With an application to precision dosing, our experimental results show that the approach can learn interpretable rules from data and may potentially lower experts' workload.
arXiv Detail & Related papers (2021-06-28T03:45:09Z) - Technology Readiness Levels for AI & ML [79.22051549519989]
Development of machine learning systems can be executed easily with modern tools, but the process is typically rushed and means-to-an-end.
Engineering systems follow well-defined processes and testing standards to streamline development for high-quality, reliable results.
We propose a proven systems engineering approach for machine learning development and deployment.
arXiv Detail & Related papers (2020-06-21T17:14:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.