Democratize with Care: The need for fairness specific features in
user-interface based open source AutoML tools
- URL: http://arxiv.org/abs/2312.12460v1
- Date: Sat, 16 Dec 2023 19:54:00 GMT
- Title: Democratize with Care: The need for fairness specific features in
user-interface based open source AutoML tools
- Authors: Sundaraparipurnan Narayanan
- Abstract summary: Automated Machine Learning (AutoML) streamlines the machine learning model development process.
This democratization allows more users (including non-experts) to access and utilize state-of-the-art machine-learning expertise.
However, AutoML tools may also propagate bias in the way these tools handle the data, model choices, and optimization approaches adopted.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: AI is increasingly playing a pivotal role in businesses and organizations,
impacting the outcomes and interests of human users. Automated Machine Learning
(AutoML) streamlines the machine learning model development process by
automating repetitive tasks and making data-driven decisions, enabling even
non-experts to construct high-quality models efficiently. This democratization
allows more users (including non-experts) to access and utilize
state-of-the-art machine-learning expertise. However, AutoML tools may also
propagate bias in the way these tools handle the data, model choices, and
optimization approaches adopted. We conducted an experimental study of
User-interface-based open source AutoML tools (DataRobot, H2O Studio, Dataiku,
and Rapidminer Studio) to examine if they had features to assist users in
developing fairness-aware machine learning models. The experiments covered the
following considerations for the evaluation of features: understanding use case
context, data representation, feature relevance and sensitivity, data bias and
preprocessing techniques, data handling capabilities, training-testing split,
hyperparameter handling, and constraints, fairness-oriented model development,
explainability and ability to download and edit models by the user. The results
revealed inadequacies in features that could support in fairness-aware model
development. Further, the results also highlight the need to establish certain
essential features for promoting fairness in AutoML tools.
Related papers
- AdaptoML-UX: An Adaptive User-centered GUI-based AutoML Toolkit for Non-AI Experts and HCI Researchers [19.602247178319992]
We introduce AdaptoML-UX, an adaptive framework that incorporates automated feature engineering, machine learning, and incremental learning.
Our toolkit demonstrates the capability to adapt efficiently to diverse problem domains and datasets.
It supports model personalization through incremental learning, customizing models to individual user behaviors.
arXiv Detail & Related papers (2024-10-22T22:52:14Z) - A Multimodal Automated Interpretability Agent [63.8551718480664]
MAIA is a system that uses neural models to automate neural model understanding tasks.
We first characterize MAIA's ability to describe (neuron-level) features in learned representations of images.
We then show that MAIA can aid in two additional interpretability tasks: reducing sensitivity to spurious features, and automatically identifying inputs likely to be mis-classified.
arXiv Detail & Related papers (2024-04-22T17:55:11Z) - The Frontier of Data Erasure: Machine Unlearning for Large Language Models [56.26002631481726]
Large Language Models (LLMs) are foundational to AI advancements.
LLMs pose risks by potentially memorizing and disseminating sensitive, biased, or copyrighted information.
Machine unlearning emerges as a cutting-edge solution to mitigate these concerns.
arXiv Detail & Related papers (2024-03-23T09:26:15Z) - CLASSify: A Web-Based Tool for Machine Learning [0.0]
This article presents an automated tool for machine learning classification problems to simplify the process of training models and producing results while providing informative visualizations and insights into the data.
We present CLASSify, an open-source tool for solving classification problems without the need for knowledge of machine learning.
arXiv Detail & Related papers (2023-10-05T15:51:36Z) - Self-directed Machine Learning [86.3709575146414]
In education science, self-directed learning has been shown to be more effective than passive teacher-guided learning.
We introduce the principal concept of Self-directed Machine Learning (SDML) and propose a framework for SDML.
Our proposed SDML process benefits from self task selection, self data selection, self model selection, self optimization strategy selection and self evaluation metric selection.
arXiv Detail & Related papers (2022-01-04T18:32:06Z) - SOLIS -- The MLOps journey from data acquisition to actionable insights [62.997667081978825]
In this paper we present a unified deployment pipeline and freedom-to-operate approach that supports all requirements while using basic cross-platform tensor framework and script language engines.
This approach however does not supply the needed procedures and pipelines for the actual deployment of machine learning capabilities in real production grade systems.
arXiv Detail & Related papers (2021-12-22T14:45:37Z) - Towards Model-informed Precision Dosing with Expert-in-the-loop Machine
Learning [0.0]
We consider a ML framework that may accelerate model learning and improve its interpretability by incorporating human experts into the model learning loop.
We propose a novel human-in-the-loop ML framework aimed at dealing with learning problems that the cost of data annotation is high.
With an application to precision dosing, our experimental results show that the approach can learn interpretable rules from data and may potentially lower experts' workload.
arXiv Detail & Related papers (2021-06-28T03:45:09Z) - Automated Machine Learning Techniques for Data Streams [91.3755431537592]
This paper surveys the state-of-the-art open-source AutoML tools, applies them to data collected from streams, and measures how their performance changes over time.
The results show that off-the-shelf AutoML tools can provide satisfactory results but in the presence of concept drift, detection or adaptation techniques have to be applied to maintain the predictive accuracy over time.
arXiv Detail & Related papers (2021-06-14T11:42:46Z) - Transfer Learning without Knowing: Reprogramming Black-box Machine
Learning Models with Scarce Data and Limited Resources [78.72922528736011]
We propose a novel approach, black-box adversarial reprogramming (BAR), that repurposes a well-trained black-box machine learning model.
Using zeroth order optimization and multi-label mapping techniques, BAR can reprogram a black-box ML model solely based on its input-output responses.
BAR outperforms state-of-the-art methods and yields comparable performance to the vanilla adversarial reprogramming method.
arXiv Detail & Related papers (2020-07-17T01:52:34Z) - Trust in AutoML: Exploring Information Needs for Establishing Trust in
Automated Machine Learning Systems [30.385703521998014]
We report results from three studies to understand the information needs of data scientists for establishing trust in AutoML systems.
We find that model performance metrics and visualizations are the most important information to data scientists when establishing their trust with an AutoML tool.
arXiv Detail & Related papers (2020-01-17T19:50:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.