Realization of Causal Representation Learning to Adjust Confounding Bias
in Latent Space
- URL: http://arxiv.org/abs/2211.08573v9
- Date: Fri, 22 Sep 2023 23:57:55 GMT
- Title: Realization of Causal Representation Learning to Adjust Confounding Bias
in Latent Space
- Authors: Jia Li, Xiang Li, Xiaowei Jia, Michael Steinbach, Vipin Kumar
- Abstract summary: Causal DAGs(Directed Acyclic Graphs) are usually considered in a 2D plane.
In this paper, we redefine causal DAG as emphdo-DAG, in which variables' values are no longer time-stamp-dependent, and timelines can be seen as axes.
- Score: 28.133104562449212
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Causal DAGs(Directed Acyclic Graphs) are usually considered in a 2D plane.
Edges indicate causal effects' directions and imply their corresponding
time-passings. Due to the natural restriction of statistical models, effect
estimation is usually approximated by averaging the individuals' correlations,
i.e., observational changes over a specific time. However, in the context of
Machine Learning on large-scale questions with complex DAGs, such slight biases
can snowball to distort global models - More importantly, it has practically
impeded the development of AI, for instance, the weak generalizability of
causal models. In this paper, we redefine causal DAG as \emph{do-DAG}, in which
variables' values are no longer time-stamp-dependent, and timelines can be seen
as axes. By geometric explanation of multi-dimensional do-DAG, we identify the
\emph{Causal Representation Bias} and its necessary factors, differentiated
from common confounding biases. Accordingly, a DL(Deep Learning)-based
framework will be proposed as the general solution, along with a realization
method and experiments to verify its feasibility.
Related papers
- Learning Directed Acyclic Graphs from Partial Orderings [9.387234607473054]
directed acyclic graphs (DAGs) are commonly used to model causal relationships among random variables.
In this paper, we consider the intermediate problem of learning DAGs when a partial causal ordering of variables is available.
We propose a general estimation framework for leveraging the partial ordering and present efficient estimation algorithms for low- and high-dimensional problems.
arXiv Detail & Related papers (2024-03-24T06:14:50Z) - A U-turn on Double Descent: Rethinking Parameter Counting in Statistical
Learning [68.76846801719095]
We show that double descent appears exactly when and where it occurs, and that its location is not inherently tied to the threshold p=n.
This provides a resolution to tensions between double descent and statistical intuition.
arXiv Detail & Related papers (2023-10-29T12:05:39Z) - It's an Alignment, Not a Trade-off: Revisiting Bias and Variance in Deep
Models [51.66015254740692]
We show that for an ensemble of deep learning based classification models, bias and variance are emphaligned at a sample level.
We study this phenomenon from two theoretical perspectives: calibration and neural collapse.
arXiv Detail & Related papers (2023-10-13T17:06:34Z) - Effect Identification in Cluster Causal Diagrams [51.42809552422494]
We introduce a new type of graphical model called cluster causal diagrams (for short, C-DAGs)
C-DAGs allow for the partial specification of relationships among variables based on limited prior knowledge.
We develop the foundations and machinery for valid causal inferences over C-DAGs.
arXiv Detail & Related papers (2022-02-22T21:27:31Z) - Variational Causal Networks: Approximate Bayesian Inference over Causal
Structures [132.74509389517203]
We introduce a parametric variational family modelled by an autoregressive distribution over the space of discrete DAGs.
In experiments, we demonstrate that the proposed variational posterior is able to provide a good approximation of the true posterior.
arXiv Detail & Related papers (2021-06-14T17:52:49Z) - Optimization Variance: Exploring Generalization Properties of DNNs [83.78477167211315]
The test error of a deep neural network (DNN) often demonstrates double descent.
We propose a novel metric, optimization variance (OV), to measure the diversity of model updates.
arXiv Detail & Related papers (2021-06-03T09:34:17Z) - Causal Expectation-Maximisation [70.45873402967297]
We show that causal inference is NP-hard even in models characterised by polytree-shaped graphs.
We introduce the causal EM algorithm to reconstruct the uncertainty about the latent variables from data about categorical manifest variables.
We argue that there appears to be an unnoticed limitation to the trending idea that counterfactual bounds can often be computed without knowledge of the structural equations.
arXiv Detail & Related papers (2020-11-04T10:25:13Z) - Efficient Variational Bayesian Structure Learning of Dynamic Graphical
Models [19.591265962713837]
Estimating time-varying graphical models is of paramount importance in various social, financial, biological, and engineering systems.
Existing methods require extensive tuning of parameters that control the graph sparsity and temporal smoothness.
We propose a low-complexity tuning-free Bayesian approach, named BADGE.
arXiv Detail & Related papers (2020-09-16T14:19:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.