ReCCoVER: Detecting Causal Confusion for Explainable Reinforcement
Learning
- URL: http://arxiv.org/abs/2203.11211v1
- Date: Mon, 21 Mar 2022 13:17:30 GMT
- Title: ReCCoVER: Detecting Causal Confusion for Explainable Reinforcement
Learning
- Authors: Jasmina Gajcin and Ivana Dusparic
- Abstract summary: Causal confusion refers to a phenomenon where an agent learns spurious correlations between features which might not hold across the entire state space.
We propose ReCCoVER, an algorithm which detects causal confusion in agent's reasoning before deployment.
- Score: 2.984934409689467
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite notable results in various fields over the recent years, deep
reinforcement learning (DRL) algorithms lack transparency, affecting user trust
and hindering their deployment to high-risk tasks. Causal confusion refers to a
phenomenon where an agent learns spurious correlations between features which
might not hold across the entire state space, preventing safe deployment to
real tasks where such correlations might be broken. In this work, we examine
whether an agent relies on spurious correlations in critical states, and
propose an alternative subset of features on which it should base its decisions
instead, to make it less susceptible to causal confusion. Our goal is to
increase transparency of DRL agents by exposing the influence of learned
spurious correlations on its decisions, and offering advice to developers about
feature selection in different parts of state space, to avoid causal confusion.
We propose ReCCoVER, an algorithm which detects causal confusion in agent's
reasoning before deployment, by executing its policy in alternative
environments where certain correlations between features do not hold. We
demonstrate our approach in taxi and grid world environments, where ReCCoVER
detects states in which an agent relies on spurious correlations and offers a
set of features that should be considered instead.
Related papers
- Causal Influence in Federated Edge Inference [34.487472866247586]
In this paper, we consider a setting where heterogeneous agents with connectivity are performing inference using unlabeled streaming data.
In order to overcome the uncertainty, agents cooperate with each other by exchanging their local inferences with and through a fusion center.
Various scenarios reflecting different agent participation patterns and fusion center policies are investigated.
arXiv Detail & Related papers (2024-05-02T13:06:50Z) - Causal State Distillation for Explainable Reinforcement Learning [16.998047658978482]
Reinforcement learning (RL) is a powerful technique for training intelligent agents, but understanding why these agents make specific decisions can be challenging.
Various approaches have been explored to address this problem, with one promising avenue being reward decomposition (RD)
RD is appealing as it sidesteps some of the concerns associated with other methods that attempt to rationalize an agent's behaviour in a post-hoc manner.
We present an extension of RD that goes beyond sub-rewards to provide more informative explanations.
arXiv Detail & Related papers (2023-12-30T00:01:22Z) - Causal Feature Selection via Transfer Entropy [59.999594949050596]
Causal discovery aims to identify causal relationships between features with observational data.
We introduce a new causal feature selection approach that relies on the forward and backward feature selection procedures.
We provide theoretical guarantees on the regression and classification errors for both the exact and the finite-sample cases.
arXiv Detail & Related papers (2023-10-17T08:04:45Z) - Learning Generalizable Agents via Saliency-Guided Features Decorrelation [25.19044461705711]
We propose Saliency-Guided Features Decorrelation to eliminate correlations between features and decisions.
RFF is utilized to estimate the complex non-linear correlations in high-dimensional images, while the saliency map is designed to identify the changed features.
Under the guidance of the saliency map, SGFD employs sample reweighting to minimize the estimated correlations related to changed features.
arXiv Detail & Related papers (2023-10-08T09:24:43Z) - Conditional Mutual Information for Disentangled Representations in
Reinforcement Learning [13.450394764597663]
Reinforcement Learning environments can produce training data with spurious correlations between features.
Disentangled representations can improve robustness, but existing disentanglement techniques that minimise mutual information between features require independent features.
We propose an auxiliary task for RL algorithms that learns a disentangled representation of high-dimensional observations with correlated features.
arXiv Detail & Related papers (2023-05-23T14:56:19Z) - Causal Disentangled Variational Auto-Encoder for Preference
Understanding in Recommendation [50.93536377097659]
This paper introduces the Causal Disentangled Variational Auto-Encoder (CaD-VAE), a novel approach for learning causal disentangled representations from interaction data in recommender systems.
The approach utilizes structural causal models to generate causal representations that describe the causal relationship between latent factors.
arXiv Detail & Related papers (2023-04-17T00:10:56Z) - Compressed Regression over Adaptive Networks [58.79251288443156]
We derive the performance achievable by a network of distributed agents that solve, adaptively and in the presence of communication constraints, a regression problem.
We devise an optimized allocation strategy where the parameters necessary for the optimization can be learned online by the agents.
arXiv Detail & Related papers (2023-04-07T13:41:08Z) - PS-ARM: An End-to-End Attention-aware Relation Mixer Network for Person
Search [56.02761592710612]
We propose a novel attention-aware relation mixer (ARM) for module person search.
Our ARM module is native and does not rely on fine-grained supervision or topological assumptions.
Our PS-ARM achieves state-of-the-art performance on both datasets.
arXiv Detail & Related papers (2022-10-07T10:04:12Z) - Robust Event-Driven Interactions in Cooperative Multi-Agent Learning [0.0]
We present an approach to reduce the communication required between agents in a Multi-Agent learning system by exploiting the inherent robustness of the underlying Markov Decision Process.
We compute so-called robustness surrogate functions (off-line), that give agents a conservative indication of how far their state measurements can deviate before they need to update other agents in the system.
This results in fully distributed decision functions, enabling agents to decide when it is necessary to update others.
arXiv Detail & Related papers (2022-04-07T11:00:39Z) - Distributed Adaptive Learning Under Communication Constraints [54.22472738551687]
This work examines adaptive distributed learning strategies designed to operate under communication constraints.
We consider a network of agents that must solve an online optimization problem from continual observation of streaming data.
arXiv Detail & Related papers (2021-12-03T19:23:48Z) - Maximizing Information Gain in Partially Observable Environments via
Prediction Reward [64.24528565312463]
This paper tackles the challenge of using belief-based rewards for a deep RL agent.
We derive the exact error between negative entropy and the expected prediction reward.
This insight provides theoretical motivation for several fields using prediction rewards.
arXiv Detail & Related papers (2020-05-11T08:13:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.