Investigating the Robustness of Artificial Intelligent Algorithms with
Mixture Experiments
- URL: http://arxiv.org/abs/2010.15551v1
- Date: Sat, 10 Oct 2020 15:38:53 GMT
- Title: Investigating the Robustness of Artificial Intelligent Algorithms with
Mixture Experiments
- Authors: Jiayi Lian, Laura Freeman, Yili Hong, and Xinwei Deng
- Abstract summary: robustness of AI algorithms is of great interest as inaccurate prediction could result in safety concerns and limit the adoption of AI systems.
A robust classification algorithm is expected to have high accuracy and low variability under different application scenarios.
We conduct a comprehensive set of mixture experiments to collect prediction performance results.
Then statistical analyses are conducted to understand how various factors affect the robustness of AI classification algorithms.
- Score: 1.877815076482061
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Artificial intelligent (AI) algorithms, such as deep learning and XGboost,
are used in numerous applications including computer vision, autonomous
driving, and medical diagnostics. The robustness of these AI algorithms is of
great interest as inaccurate prediction could result in safety concerns and
limit the adoption of AI systems. In this paper, we propose a framework based
on design of experiments to systematically investigate the robustness of AI
classification algorithms. A robust classification algorithm is expected to
have high accuracy and low variability under different application scenarios.
The robustness can be affected by a wide range of factors such as the imbalance
of class labels in the training dataset, the chosen prediction algorithm, the
chosen dataset of the application, and a change of distribution in the training
and test datasets. To investigate the robustness of AI classification
algorithms, we conduct a comprehensive set of mixture experiments to collect
prediction performance results. Then statistical analyses are conducted to
understand how various factors affect the robustness of AI classification
algorithms. We summarize our findings and provide suggestions to practitioners
in AI applications.
Related papers
- Enhancing Feature Selection and Interpretability in AI Regression Tasks Through Feature Attribution [38.53065398127086]
This study investigates the potential of feature attribution methods to filter out uninformative features in input data for regression problems.
We introduce a feature selection pipeline that combines Integrated Gradients with k-means clustering to select an optimal set of variables from the initial data space.
To validate the effectiveness of this approach, we apply it to a real-world industrial problem - blade vibration analysis in the development process of turbo machinery.
arXiv Detail & Related papers (2024-09-25T09:50:51Z) - Multi-Dimensional Ability Diagnosis for Machine Learning Algorithms [88.93372675846123]
We propose a task-agnostic evaluation framework Camilla for evaluating machine learning algorithms.
We use cognitive diagnosis assumptions and neural networks to learn the complex interactions among algorithms, samples and the skills of each sample.
In our experiments, Camilla outperforms state-of-the-art baselines on the metric reliability, rank consistency and rank stability.
arXiv Detail & Related papers (2023-07-14T03:15:56Z) - Human-Centric Multimodal Machine Learning: Recent Advances and Testbed
on AI-based Recruitment [66.91538273487379]
There is a certain consensus about the need to develop AI applications with a Human-Centric approach.
Human-Centric Machine Learning needs to be developed based on four main requirements: (i) utility and social good; (ii) privacy and data ownership; (iii) transparency and accountability; and (iv) fairness in AI-driven decision-making processes.
We study how current multimodal algorithms based on heterogeneous sources of information are affected by sensitive elements and inner biases in the data.
arXiv Detail & Related papers (2023-02-13T16:44:44Z) - Do-AIQ: A Design-of-Experiment Approach to Quality Evaluation of AI
Mislabel Detection Algorithm [0.0]
The quality of Artificial Intelligence (AI) algorithms is of significant importance for confidently adopting algorithms in various applications such as cybersecurity, healthcare, and autonomous driving.
This work presents a principled framework of using a design-of-experimental approach to systematically evaluate the quality of AI algorithms, named as Do-AIQ.
arXiv Detail & Related papers (2022-08-21T19:47:41Z) - Correntropy-Based Logistic Regression with Automatic Relevance
Determination for Robust Sparse Brain Activity Decoding [18.327196310636864]
We introduce the correntropy learning framework into the automatic relevance determination based sparse classification model.
We evaluate it on a synthetic dataset, an electroencephalogram (EEG) dataset, and a functional magnetic resonance imaging (fMRI) dataset.
arXiv Detail & Related papers (2022-07-20T06:49:23Z) - Towards Diverse Evaluation of Class Incremental Learning: A Representation Learning Perspective [67.45111837188685]
Class incremental learning (CIL) algorithms aim to continually learn new object classes from incrementally arriving data.
We experimentally analyze neural network models trained by CIL algorithms using various evaluation protocols in representation learning.
arXiv Detail & Related papers (2022-06-16T11:44:11Z) - RL-EA: A Reinforcement Learning-Based Evolutionary Algorithm Framework
for Electromagnetic Detection Satellite Scheduling Problem [6.438148195340613]
This paper proposes a mixed-integer programming model for the EDSSP problem and an evolutionary algorithm framework based on reinforcement learning (RL-EA)
Various scales experiments are used to examine the planning effect of the proposed algorithm.
arXiv Detail & Related papers (2022-06-12T08:53:56Z) - Robustification of Online Graph Exploration Methods [59.50307752165016]
We study a learning-augmented variant of the classical, notoriously hard online graph exploration problem.
We propose an algorithm that naturally integrates predictions into the well-known Nearest Neighbor (NN) algorithm.
arXiv Detail & Related papers (2021-12-10T10:02:31Z) - FairCVtest Demo: Understanding Bias in Multimodal Learning with a
Testbed in Fair Automatic Recruitment [79.23531577235887]
This demo shows the capacity of the Artificial Intelligence (AI) behind a recruitment tool to extract sensitive information from unstructured data.
Aditionally, the demo includes a new algorithm for discrimination-aware learning which eliminates sensitive information in our multimodal AI framework.
arXiv Detail & Related papers (2020-09-12T17:45:09Z) - Bias in Multimodal AI: Testbed for Fair Automatic Recruitment [73.85525896663371]
We study how current multimodal algorithms based on heterogeneous sources of information are affected by sensitive elements and inner biases in the data.
We train automatic recruitment algorithms using a set of multimodal synthetic profiles consciously scored with gender and racial biases.
Our methodology and results show how to generate fairer AI-based tools in general, and in particular fairer automated recruitment systems.
arXiv Detail & Related papers (2020-04-15T15:58:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.