Reinforcement Learning of Causal Variables Using Mediation Analysis
- URL: http://arxiv.org/abs/2010.15745v2
- Date: Tue, 17 May 2022 10:56:48 GMT
- Title: Reinforcement Learning of Causal Variables Using Mediation Analysis
- Authors: Tue Herlau, Rasmus Larsen
- Abstract summary: We consider the problem of building a general reinforcement learning agent which uses experience to construct a causal graph of the environment.
We show the method can learn a plausible causal graph in a grid-world environment, and the agent obtains an improvement in performance when using the causally informed policy.
- Score: 0.15229257192293197
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many open problems in machine learning are intrinsically related to
causality, however, the use of causal analysis in machine learning is still in
its early stage. Within a general reinforcement learning setting, we consider
the problem of building a general reinforcement learning agent which uses
experience to construct a causal graph of the environment, and use this graph
to inform its policy. Our approach has three characteristics: First, we learn a
simple, coarse-grained causal graph, in which the variables reflect states at
many time instances, and the interventions happen at the level of policies,
rather than individual actions. Secondly, we use mediation analysis to obtain
an optimization target. By minimizing this target, we define the causal
variables. Thirdly, our approach relies on estimating conditional expectations
rather the familiar expected return from reinforcement learning, and we
therefore apply a generalization of Bellman's equations. We show the method can
learn a plausible causal graph in a grid-world environment, and the agent
obtains an improvement in performance when using the causally informed policy.
To our knowledge, this is the first attempt to apply causal analysis in a
reinforcement learning setting without strict restrictions on the number of
states. We have observed that mediation analysis provides a promising avenue
for transforming the problem of causal acquisition into one of cost-function
minimization, but importantly one which involves estimating conditional
expectations. This is a new challenge, and we think that causal reinforcement
learning will involve development methods suited for online estimation of such
conditional expectations. Finally, a benefit of our approach is the use of very
simple causal models, which are arguably a more natural model of human causal
understanding.
Related papers
- Learning by Doing: An Online Causal Reinforcement Learning Framework
with Causal-Aware Policy [40.33036146207819]
We consider explicitly modeling the generation process of states with the graphical causal model.
We formulate the causal structure updating into the RL interaction process with active intervention learning of the environment.
arXiv Detail & Related papers (2024-02-07T14:09:34Z) - Multi-modal Causal Structure Learning and Root Cause Analysis [67.67578590390907]
We propose Mulan, a unified multi-modal causal structure learning method for root cause localization.
We leverage a log-tailored language model to facilitate log representation learning, converting log sequences into time-series data.
We also introduce a novel key performance indicator-aware attention mechanism for assessing modality reliability and co-learning a final causal graph.
arXiv Detail & Related papers (2024-02-04T05:50:38Z) - Interpretable Imitation Learning with Dynamic Causal Relations [65.18456572421702]
We propose to expose captured knowledge in the form of a directed acyclic causal graph.
We also design this causal discovery process to be state-dependent, enabling it to model the dynamics in latent causal graphs.
The proposed framework is composed of three parts: a dynamic causal discovery module, a causality encoding module, and a prediction module, and is trained in an end-to-end manner.
arXiv Detail & Related papers (2023-09-30T20:59:42Z) - Bayesian Learning for Dynamic Inference [2.2843885788439793]
In some sequential estimation problems, the future values of the quantity to be estimated depend on the estimate of its current value.
We formulate the Bayesian learning problem for dynamic inference, where the unknown quantity-generation model is assumed to be randomly drawn.
We derive the optimal Bayesian learning rules, both offline and online, to minimize the inference loss.
arXiv Detail & Related papers (2022-12-30T19:16:23Z) - A Meta-Reinforcement Learning Algorithm for Causal Discovery [3.4806267677524896]
Causal structures can enable models to go beyond pure correlation-based inference.
Finding causal structures from data poses a significant challenge both in computational effort and accuracy.
We develop a meta-reinforcement learning algorithm that performs causal discovery by learning to perform interventions.
arXiv Detail & Related papers (2022-07-18T09:26:07Z) - Evaluation Methods and Measures for Causal Learning Algorithms [33.07234268724662]
We focus on the two fundamental causal-inference tasks and causality-aware machine learning tasks.
The survey seeks to bring to the forefront the urgency of developing publicly available benchmarks and consensus-building standards for causal learning evaluation with observational data.
arXiv Detail & Related papers (2022-02-07T00:24:34Z) - Stateful Offline Contextual Policy Evaluation and Learning [88.9134799076718]
We study off-policy evaluation and learning from sequential data.
We formalize the relevant causal structure of problems such as dynamic personalized pricing.
We show improved out-of-sample policy performance in this class of relevant problems.
arXiv Detail & Related papers (2021-10-19T16:15:56Z) - Systematic Evaluation of Causal Discovery in Visual Model Based
Reinforcement Learning [76.00395335702572]
A central goal for AI and causality is the joint discovery of abstract representations and causal structure.
Existing environments for studying causal induction are poorly suited for this objective because they have complicated task-specific causal graphs.
In this work, our goal is to facilitate research in learning representations of high-level variables as well as causal structures among them.
arXiv Detail & Related papers (2021-07-02T05:44:56Z) - Constrained Learning with Non-Convex Losses [119.8736858597118]
Though learning has become a core technology of modern information processing, there is now ample evidence that it can lead to biased, unsafe, and prejudiced solutions.
arXiv Detail & Related papers (2021-03-08T23:10:33Z) - Towards Causal Representation Learning [96.110881654479]
The two fields of machine learning and graphical causality arose and developed separately.
There is now cross-pollination and increasing interest in both fields to benefit from the advances of the other.
arXiv Detail & Related papers (2021-02-22T15:26:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.