A Bandit Model for Human-Machine Decision Making with Private
Information and Opacity
- URL: http://arxiv.org/abs/2007.04800v3
- Date: Tue, 3 May 2022 17:27:15 GMT
- Title: A Bandit Model for Human-Machine Decision Making with Private
Information and Opacity
- Authors: Sebastian Bordt, Ulrike von Luxburg
- Abstract summary: We show a two-player learning problem where one player is the machine and the other the human.
A lower bound quantifies the worst-case hardness of optimally advising a decision maker who is opaque.
An upper bound shows that a simple coordination strategy is nearly minimax optimal.
- Score: 16.665883787432858
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Applications of machine learning inform human decision makers in a broad
range of tasks. The resulting problem is usually formulated in terms of a
single decision maker. We argue that it should rather be described as a
two-player learning problem where one player is the machine and the other the
human. While both players try to optimize the final decision, the setup is
often characterized by (1) the presence of private information and (2) opacity,
that is imperfect understanding between the decision makers. We prove that both
properties can complicate decision making considerably. A lower bound
quantifies the worst-case hardness of optimally advising a decision maker who
is opaque or has access to private information. An upper bound shows that a
simple coordination strategy is nearly minimax optimal. More efficient learning
is possible under certain assumptions on the problem, for example that both
players learn to take actions independently. Such assumptions are implicit in
existing literature, for example in medical applications of machine learning,
but have not been described or justified theoretically.
Related papers
- Capturing the Complexity of Human Strategic Decision-Making with Machine Learning [4.308322597847064]
We conduct the largest study to date of strategic decision-making in the context of initial play in two-player matrix games.
We show that a deep neural network trained on these data predicts people's choices better than leading theories of strategic behavior.
arXiv Detail & Related papers (2024-08-15T00:39:42Z) - Optimising Human-AI Collaboration by Learning Convincing Explanations [62.81395661556852]
We propose a method for a collaborative system that remains safe by having a human making decisions.
Ardent enables efficient and effective decision-making by adapting to individual preferences for explanations.
arXiv Detail & Related papers (2023-11-13T16:00:16Z) - Fairness Implications of Heterogeneous Treatment Effect Estimation with
Machine Learning Methods in Policy-making [0.0]
We argue that standard AI Fairness approaches for predictive machine learning are not suitable for all causal machine learning applications.
We argue that policy-making is best seen as a joint decision where the causal machine learning model usually only has indirect power.
arXiv Detail & Related papers (2023-09-02T03:06:14Z) - Learning to Guide Human Experts via Personalized Large Language Models [23.7625973884849]
In learning to defer, a predictor identifies risky decisions and defers them to a human expert.
In learning to guide, the machine provides guidance useful to guide decision-making, and the human is entirely responsible for coming up with a decision.
arXiv Detail & Related papers (2023-08-11T09:36:33Z) - Assisting Human Decisions in Document Matching [52.79491990823573]
We devise a proxy matching task that allows us to evaluate which kinds of assistive information improve decision makers' performance.
We find that providing black-box model explanations reduces users' accuracy on the matching task.
On the other hand, custom methods that are designed to closely attend to some task-specific desiderata are found to be effective in improving user performance.
arXiv Detail & Related papers (2023-02-16T17:45:20Z) - Explainable Data-Driven Optimization: From Context to Decision and Back
Again [76.84947521482631]
Data-driven optimization uses contextual information and machine learning algorithms to find solutions to decision problems with uncertain parameters.
We introduce a counterfactual explanation methodology tailored to explain solutions to data-driven problems.
We demonstrate our approach by explaining key problems in operations management such as inventory management and routing.
arXiv Detail & Related papers (2023-01-24T15:25:16Z) - The Paradox of Choice: Using Attention in Hierarchical Reinforcement
Learning [59.777127897688594]
We present an online, model-free algorithm to learn affordances that can be used to further learn subgoal options.
We investigate the role of hard versus soft attention in training data collection, abstract value learning in long-horizon tasks, and handling a growing number of choices.
arXiv Detail & Related papers (2022-01-24T13:18:02Z) - Learning the Preferences of Uncertain Humans with Inverse Decision
Theory [10.926992035470372]
We study the setting of inverse decision theory (IDT), a framework where a human is observed making non-sequential binary decisions under uncertainty.
In IDT, the human's preferences are conveyed through their loss function, which expresses a tradeoff between different types of mistakes.
We show that it is actually easier to identify preferences when the decision problem is more uncertain.
arXiv Detail & Related papers (2021-06-19T00:11:13Z) - Indecision Modeling [50.00689136829134]
It is important that AI systems act in ways which align with human values.
People are often indecisive, and especially so when their decision has moral implications.
arXiv Detail & Related papers (2020-12-15T18:32:37Z) - An Information-Theoretic Approach to Personalized Explainable Machine
Learning [92.53970625312665]
We propose a simple probabilistic model for the predictions and user knowledge.
We quantify the effect of an explanation by the conditional mutual information between the explanation and prediction.
arXiv Detail & Related papers (2020-03-01T13:06:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.