Identifying Concerns When Specifying Machine Learning-Enabled Systems: A
Perspective-Based Approach
- URL: http://arxiv.org/abs/2309.07980v1
- Date: Thu, 14 Sep 2023 18:31:16 GMT
- Title: Identifying Concerns When Specifying Machine Learning-Enabled Systems: A
Perspective-Based Approach
- Authors: Hugo Villamizar, Marcos Kalinowski, Helio Lopes, Daniel Mendez
- Abstract summary: PerSpecML is a perspective-based approach for specifying ML-enabled systems.
It helps practitioners identify which attributes, including ML and non-ML components, are important to contribute to the overall system's quality.
- Score: 1.2184324428571227
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Engineering successful machine learning (ML)-enabled systems poses various
challenges from both a theoretical and a practical side. Among those challenges
are how to effectively address unrealistic expectations of ML capabilities from
customers, managers and even other team members, and how to connect business
value to engineering and data science activities composed by interdisciplinary
teams. In this paper, we present PerSpecML, a perspective-based approach for
specifying ML-enabled systems that helps practitioners identify which
attributes, including ML and non-ML components, are important to contribute to
the overall system's quality. The approach involves analyzing 59 concerns
related to typical tasks that practitioners face in ML projects, grouping them
into five perspectives: system objectives, user experience, infrastructure,
model, and data. Together, these perspectives serve to mediate the
communication between business owners, domain experts, designers, software and
ML engineers, and data scientists. The creation of PerSpecML involved a series
of validations conducted in different contexts: (i) in academia, (ii) with
industry representatives, and (iii) in two real industrial case studies. As a
result of the diverse validations and continuous improvements, PerSpecML stands
as a promising approach, poised to positively impact the specification of
ML-enabled systems, particularly helping to reveal key components that would
have been otherwise missed without using PerSpecML.
Related papers
- Understanding the Role of LLMs in Multimodal Evaluation Benchmarks [77.59035801244278]
This paper investigates the role of the Large Language Model (LLM) backbone in Multimodal Large Language Models (MLLMs) evaluation.
Our study encompasses four diverse MLLM benchmarks and eight state-of-the-art MLLMs.
Key findings reveal that some benchmarks allow high performance even without visual inputs and up to 50% of error rates can be attributed to insufficient world knowledge in the LLM backbone.
arXiv Detail & Related papers (2024-10-16T07:49:13Z) - Towards Effective Collaboration between Software Engineers and Data Scientists developing Machine Learning-Enabled Systems [1.1153433121962064]
Development of Machine Learning (ML)-enabled systems encompasses several social and technical challenges.
This paper has the objective of understanding how to enhance the collaboration between two key actors in building these systems: software engineers and data scientists.
Our research has found that collaboration between these actors is important for effectively developing ML-enabled systems.
arXiv Detail & Related papers (2024-07-22T17:35:18Z) - AesBench: An Expert Benchmark for Multimodal Large Language Models on
Image Aesthetics Perception [64.25808552299905]
AesBench is an expert benchmark aiming to comprehensively evaluate the aesthetic perception capacities of MLLMs.
We construct an Expert-labeled Aesthetics Perception Database (EAPD), which features diversified image contents and high-quality annotations provided by professional aesthetic experts.
We propose a set of integrative criteria to measure the aesthetic perception abilities of MLLMs from four perspectives, including Perception (AesP), Empathy (AesE), Assessment (AesA) and Interpretation (AesI)
arXiv Detail & Related papers (2024-01-16T10:58:07Z) - Knowledge Plugins: Enhancing Large Language Models for Domain-Specific
Recommendations [50.81844184210381]
We propose a general paradigm that augments large language models with DOmain-specific KnowledgE to enhance their performance on practical applications, namely DOKE.
This paradigm relies on a domain knowledge extractor, working in three steps: 1) preparing effective knowledge for the task; 2) selecting the knowledge for each specific sample; and 3) expressing the knowledge in an LLM-understandable way.
arXiv Detail & Related papers (2023-11-16T07:09:38Z) - Model-driven Engineering for Machine Learning Components: A Systematic
Literature Review [8.810090413018798]
We analyzed studies with respect to several areas of interest and identified the key motivations behind using MDE4ML.
We also discuss the gaps in existing literature and provide recommendations for future work.
arXiv Detail & Related papers (2023-11-01T04:29:47Z) - Machine Learning-Enabled Software and System Architecture Frameworks [48.87872564630711]
The stakeholders with data science and Machine Learning related concerns, such as data scientists and data engineers, are yet to be included in existing architecture frameworks.
We surveyed 61 subject matter experts from over 25 organizations in 10 countries.
arXiv Detail & Related papers (2023-08-09T21:54:34Z) - Towards Perspective-Based Specification of Machine Learning-Enabled
Systems [1.3406258114080236]
This paper describes our work towards a perspective-based approach for specifying ML-enabled systems.
The approach involves analyzing a set of 45 ML concerns grouped into five perspectives: objectives, user experience, infrastructure, model, and data.
The main contribution of this paper is to provide two new artifacts that can be used to help specifying ML-enabled systems.
arXiv Detail & Related papers (2022-06-20T13:09:23Z) - Machine Learning Operations (MLOps): Overview, Definition, and
Architecture [0.0]
The paradigm of Machine Learning Operations (MLOps) addresses this issue.
MLOps is still a vague term and its consequences for researchers and professionals are ambiguous.
We provide an aggregated overview of the necessary components, and roles, as well as the associated architecture and principles.
arXiv Detail & Related papers (2022-05-04T19:38:48Z) - Panoramic Learning with A Standardized Machine Learning Formalism [116.34627789412102]
This paper presents a standardized equation of the learning objective, that offers a unifying understanding of diverse ML algorithms.
It also provides guidance for mechanic design of new ML solutions, and serves as a promising vehicle towards panoramic learning with all experiences.
arXiv Detail & Related papers (2021-08-17T17:44:38Z) - Characterizing and Detecting Mismatch in Machine-Learning-Enabled
Systems [1.4695979686066065]
Development and deployment of machine learning systems remains a challenge.
In this paper, we report our findings and their implications for improving end-to-end ML-enabled system development.
arXiv Detail & Related papers (2021-03-25T19:40:29Z) - Understanding the Usability Challenges of Machine Learning In
High-Stakes Decision Making [67.72855777115772]
Machine learning (ML) is being applied to a diverse and ever-growing set of domains.
In many cases, domain experts -- who often have no expertise in ML or data science -- are asked to use ML predictions to make high-stakes decisions.
We investigate the ML usability challenges present in the domain of child welfare screening through a series of collaborations with child welfare screeners.
arXiv Detail & Related papers (2021-03-02T22:50:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.