A Dynamical View of the Question of Why
- URL: http://arxiv.org/abs/2402.10240v2
- Date: Tue, 27 Feb 2024 22:11:42 GMT
- Title: A Dynamical View of the Question of Why
- Authors: Mehdi Fatemi and Sindhu Gowda
- Abstract summary: We present two key lemmas to compute causal contributions and frame them as reinforcement learning problems.
Our approach offers formal and computational tools for uncovering and quantifying causal relationships in diffusion processes.
In fairly intricate experiments and through sheer learning, our framework reveals and quantifies causal links, which otherwise seem inexplicable.
- Score: 10.243488468625783
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We address causal reasoning in multivariate time series data generated by
stochastic processes. Existing approaches are largely restricted to static
settings, ignoring the continuity and emission of variations across time. In
contrast, we propose a learning paradigm that directly establishes causation
between events in the course of time. We present two key lemmas to compute
causal contributions and frame them as reinforcement learning problems. Our
approach offers formal and computational tools for uncovering and quantifying
causal relationships in diffusion processes, subsuming various important
settings such as discrete-time Markov decision processes. Finally, in fairly
intricate experiments and through sheer learning, our framework reveals and
quantifies causal links, which otherwise seem inexplicable.
Related papers
- Decoupled Marked Temporal Point Process using Neural Ordinary Differential Equations [14.828081841581296]
A Marked Temporal Point Process (MTPP) is a process whose realization is a set of event-time data.
Recent studies have utilized deep neural networks to capture complex temporal dependencies of events.
We propose a Decoupled MTPP framework that disentangles characterization of a process into a set of evolving influences from different events.
arXiv Detail & Related papers (2024-06-10T10:15:32Z) - On the Identification of Temporally Causal Representation with Instantaneous Dependence [50.14432597910128]
Temporally causal representation learning aims to identify the latent causal process from time series observations.
Most methods require the assumption that the latent causal processes do not have instantaneous relations.
We propose an textbfIDentification framework for instantanetextbfOus textbfLatent dynamics.
arXiv Detail & Related papers (2024-05-24T08:08:05Z) - Multi-modal Causal Structure Learning and Root Cause Analysis [67.67578590390907]
We propose Mulan, a unified multi-modal causal structure learning method for root cause localization.
We leverage a log-tailored language model to facilitate log representation learning, converting log sequences into time-series data.
We also introduce a novel key performance indicator-aware attention mechanism for assessing modality reliability and co-learning a final causal graph.
arXiv Detail & Related papers (2024-02-04T05:50:38Z) - Causal Discovery and Prediction: Methods and Algorithms [0.0]
In this thesis we introduce a generic a-priori assessment of each possible intervention.
We propose an active learning algorithm that identifies the causal relations in any given causal model.
arXiv Detail & Related papers (2023-09-18T01:19:37Z) - BayesIMP: Uncertainty Quantification for Causal Data Fusion [52.184885680729224]
We study the causal data fusion problem, where datasets pertaining to multiple causal graphs are combined to estimate the average treatment effect of a target variable.
We introduce a framework which combines ideas from probabilistic integration and kernel mean embeddings to represent interventional distributions in the reproducing kernel Hilbert space.
arXiv Detail & Related papers (2021-06-07T10:14:18Z) - Consistency of mechanistic causal discovery in continuous-time using
Neural ODEs [85.7910042199734]
We consider causal discovery in continuous-time for the study of dynamical systems.
We propose a causal discovery algorithm based on penalized Neural ODEs.
arXiv Detail & Related papers (2021-05-06T08:48:02Z) - Contrastive learning of strong-mixing continuous-time stochastic
processes [53.82893653745542]
Contrastive learning is a family of self-supervised methods where a model is trained to solve a classification task constructed from unlabeled data.
We show that a properly constructed contrastive learning task can be used to estimate the transition kernel for small-to-mid-range intervals in the diffusion case.
arXiv Detail & Related papers (2021-03-03T23:06:47Z) - Disentangling Observed Causal Effects from Latent Confounders using
Method of Moments [67.27068846108047]
We provide guarantees on identifiability and learnability under mild assumptions.
We develop efficient algorithms based on coupled tensor decomposition with linear constraints to obtain scalable and guaranteed solutions.
arXiv Detail & Related papers (2021-01-17T07:48:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.