HEX: Human-in-the-loop Explainability via Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2206.01343v1
- Date: Thu, 2 Jun 2022 23:53:40 GMT
- Title: HEX: Human-in-the-loop Explainability via Deep Reinforcement Learning
- Authors: Michael T. Lash
- Abstract summary: We propose HEX, a human-in-the-loop deep reinforcement learning approach to machine learning explainability (MLX)
Our formulation explicitly considers the decision boundary of the ML model in question, rather than the underlying training data.
Our proposed methods thus synthesize HITL MLX policies that explicitly capture the decision boundary of the model in question for use in limited data scenarios.
- Score: 2.322461721824713
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The use of machine learning (ML) models in decision-making contexts,
particularly those used in high-stakes decision-making, are fraught with issue
and peril since a person - not a machine - must ultimately be held accountable
for the consequences of the decisions made using such systems. Machine learning
explainability (MLX) promises to provide decision-makers with
prediction-specific rationale, assuring them that the model-elicited
predictions are made for the right reasons and are thus reliable. Few works
explicitly consider this key human-in-the-loop (HITL) component, however. In
this work we propose HEX, a human-in-the-loop deep reinforcement learning
approach to MLX. HEX incorporates 0-distrust projection to synthesize decider
specific explanation-providing policies from any arbitrary classification
model. HEX is also constructed to operate in limited or reduced training data
scenarios, such as those employing federated learning. Our formulation
explicitly considers the decision boundary of the ML model in question, rather
than the underlying training data, which is a shortcoming of many
model-agnostic MLX methods. Our proposed methods thus synthesize HITL MLX
policies that explicitly capture the decision boundary of the model in question
for use in limited data scenarios.
Related papers
- Self-Healing Machine Learning: A Framework for Autonomous Adaptation in Real-World Environments [50.310636905746975]
Real-world machine learning systems often encounter model performance degradation due to distributional shifts in the underlying data generating process.
Existing approaches to addressing shifts, such as concept drift adaptation, are limited by their reason-agnostic nature.
We propose self-healing machine learning (SHML) to overcome these limitations.
arXiv Detail & Related papers (2024-10-31T20:05:51Z) - Modeling Boundedly Rational Agents with Latent Inference Budgets [56.24971011281947]
We introduce a latent inference budget model (L-IBM) that models agents' computational constraints explicitly.
L-IBMs make it possible to learn agent models using data from diverse populations of suboptimal actors.
We show that L-IBMs match or outperform Boltzmann models of decision-making under uncertainty.
arXiv Detail & Related papers (2023-12-07T03:55:51Z) - Leaving the Nest: Going Beyond Local Loss Functions for
Predict-Then-Optimize [57.22851616806617]
We show that our method achieves state-of-the-art results in four domains from the literature.
Our approach outperforms the best existing method by nearly 200% when the localness assumption is broken.
arXiv Detail & Related papers (2023-05-26T11:17:45Z) - A Meta-heuristic Approach to Estimate and Explain Classifier Uncertainty [0.4264192013842096]
This work proposes a set of class-independent meta-heuristics that can characterize the complexity of an instance in terms of factors are mutually relevant to both human and machine learning decision-making.
The proposed measures and framework hold promise for improving model development for more complex instances, as well as providing a new means of model abstention and explanation.
arXiv Detail & Related papers (2023-04-20T13:09:28Z) - Use Perturbations when Learning from Explanations [51.19736333434313]
Machine learning from explanations (MLX) is an approach to learning that uses human-provided explanations of relevant or irrelevant features for each input.
We recast MLX as a problem, where human explanations specify a lower dimensional manifold from which robustnesss can be drawn.
We consider various approaches to achieving robustness, leading to improved performance over prior MLX methods.
arXiv Detail & Related papers (2023-03-11T14:57:52Z) - Prescriptive Machine Learning for Automated Decision Making: Challenges
and Opportunities [0.0]
Recent applications of machine learning (ML) reveal a noticeable shift from its use for predictive modeling to its use for prescriptive modeling.
prescriptive modeling comes with new technical conditions for learning and new demands regarding reliability, responsibility, and the ethics of decision making.
To support the data-driven design of decision-making agents that act in a rational but at the same time responsible manner, a rigorous methodological foundation of prescriptive ML is needed.
arXiv Detail & Related papers (2021-12-15T17:02:07Z) - SMACE: A New Method for the Interpretability of Composite Decision
Systems [3.1542695050861544]
We present SMACE, Semi-Model-Agnostic Contextual Explainer, a novel interpretability method for decision systems.
We show that established model-agnostic approaches produce poor results in this framework.
arXiv Detail & Related papers (2021-11-16T19:37:35Z) - Leveraging Expert Consistency to Improve Algorithmic Decision Support [62.61153549123407]
We explore the use of historical expert decisions as a rich source of information that can be combined with observed outcomes to narrow the construct gap.
We propose an influence function-based methodology to estimate expert consistency indirectly when each case in the data is assessed by a single expert.
Our empirical evaluation, using simulations in a clinical setting and real-world data from the child welfare domain, indicates that the proposed approach successfully narrows the construct gap.
arXiv Detail & Related papers (2021-01-24T05:40:29Z) - Identification of Unexpected Decisions in Partially Observable
Monte-Carlo Planning: a Rule-Based Approach [78.05638156687343]
We propose a methodology for analyzing POMCP policies by inspecting their traces.
The proposed method explores local properties of policy behavior to identify unexpected decisions.
We evaluate our approach on Tiger, a standard benchmark for POMDPs, and a real-world problem related to mobile robot navigation.
arXiv Detail & Related papers (2020-12-23T15:09:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.