Multi-Agent Algorithmic Recourse
- URL: http://arxiv.org/abs/2110.00673v1
- Date: Fri, 1 Oct 2021 22:54:47 GMT
- Title: Multi-Agent Algorithmic Recourse
- Authors: Andrew O'Brien, Edward Kim
- Abstract summary: We show that when the assumption of a single agent environment is relaxed, current approaches to algorithmic recourse fail to guarantee certain ethically desirable properties.
We propose a new game theory inspired framework for providing algorithmic recourse in a multi-agent environment.
- Score: 7.23389716633927
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The recent adoption of machine learning as a tool in real world decision
making has spurred interest in understanding how these decisions are being
made. Counterfactual Explanations are a popular interpretable machine learning
technique that aims to understand how a machine learning model would behave if
given alternative inputs. Many explanations attempt to go further and recommend
actions an individual could take to obtain a more desirable output from the
model. These recommendations are known as algorithmic recourse. Past work has
largely focused on the effect algorithmic recourse has on a single agent. In
this work, we show that when the assumption of a single agent environment is
relaxed, current approaches to algorithmic recourse fail to guarantee certain
ethically desirable properties. Instead, we propose a new game theory inspired
framework for providing algorithmic recourse in a multi-agent environment that
does guarantee these properties.
Related papers
- VCNet: A self-explaining model for realistic counterfactual generation [52.77024349608834]
Counterfactual explanation is a class of methods to make local explanations of machine learning decisions.
We present VCNet-Variational Counter Net, a model architecture that combines a predictor and a counterfactual generator.
We show that VCNet is able to both generate predictions, and to generate counterfactual explanations without having to solve another minimisation problem.
arXiv Detail & Related papers (2022-12-21T08:45:32Z) - Inapplicable Actions Learning for Knowledge Transfer in Reinforcement
Learning [3.194414753332705]
We show that learning inapplicable actions greatly improves the sample efficiency of RL algorithms.
Thanks to the transferability of the knowledge acquired, it can be reused in other tasks and domains to make the learning process more efficient.
arXiv Detail & Related papers (2022-11-28T17:45:39Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Emergent Instabilities in Algorithmic Feedback Loops [3.4711828357576855]
We explore algorithmic confounding in recommendation algorithms through teacher-student learning simulations.
Results highlight the need to account for emergent behaviors from interactions between people and algorithms.
arXiv Detail & Related papers (2022-01-18T18:58:03Z) - Synthesizing explainable counterfactual policies for algorithmic
recourse with program synthesis [18.485744170172545]
We learn a program that outputs a sequence of explainable counterfactual actions given a user description and a causal graph.
An experimental evaluation on synthetic and real-world datasets shows how our approach generates effective interventions.
arXiv Detail & Related papers (2022-01-18T17:16:45Z) - An actor-critic algorithm with policy gradients to solve the job shop
scheduling problem using deep double recurrent agents [1.3812010983144802]
We propose a deep reinforcement learning methodology for the job shop scheduling problem (JSSP)
The aim is to build up a greedy-like able to learn on some distribution of JSSP instances, different in the number of jobs and machines.
As expected, the model can generalize, to some extent, to larger problems or instances originated by a different distribution from the one used in training.
arXiv Detail & Related papers (2021-10-18T07:55:39Z) - Robust Predictable Control [149.71263296079388]
We show that our method achieves much tighter compression than prior methods, achieving up to 5x higher reward than a standard information bottleneck.
We also demonstrate that our method learns policies that are more robust and generalize better to new tasks.
arXiv Detail & Related papers (2021-09-07T17:29:34Z) - Modularity in Reinforcement Learning via Algorithmic Independence in
Credit Assignment [79.5678820246642]
We show that certain action-value methods are more sample efficient than policy-gradient methods on transfer problems that require only sparse changes to a sequence of previously optimal decisions.
We generalize the recently proposed societal decision-making framework as a more granular formalism than the Markov decision process.
arXiv Detail & Related papers (2021-06-28T21:29:13Z) - Model-Agnostic Explanations using Minimal Forcing Subsets [11.420687735660097]
We propose a new model-agnostic algorithm to identify a minimal set of training samples that are indispensable for a given model's decision.
Our algorithm identifies such a set of "indispensable" samples iteratively by solving a constrained optimization problem.
Results show that our algorithm is an effective and easy-to-comprehend tool that helps to better understand local model behavior.
arXiv Detail & Related papers (2020-11-01T22:45:16Z) - Discovering Reinforcement Learning Algorithms [53.72358280495428]
Reinforcement learning algorithms update an agent's parameters according to one of several possible rules.
This paper introduces a new meta-learning approach that discovers an entire update rule.
It includes both 'what to predict' (e.g. value functions) and 'how to learn from it' by interacting with a set of environments.
arXiv Detail & Related papers (2020-07-17T07:38:39Z) - Plausible Counterfactuals: Auditing Deep Learning Classifiers with
Realistic Adversarial Examples [84.8370546614042]
Black-box nature of Deep Learning models has posed unanswered questions about what they learn from data.
Generative Adversarial Network (GAN) and multi-objectives are used to furnish a plausible attack to the audited model.
Its utility is showcased within a human face classification task, unveiling the enormous potential of the proposed framework.
arXiv Detail & Related papers (2020-03-25T11:08:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.