Almost Optimal Universal Lower Bound for Learning Causal DAGs with
Atomic Interventions
- URL: http://arxiv.org/abs/2111.05070v1
- Date: Tue, 9 Nov 2021 11:58:44 GMT
- Title: Almost Optimal Universal Lower Bound for Learning Causal DAGs with
Atomic Interventions
- Authors: Vibhor Porwal, Piyush Srivastava, Gaurav Sinha
- Abstract summary: We prove a new universal lower bound on the number of atomic interventions that an algorithm would need to perform in order to orient a given MEC.
Our lower bound is provably better than previously known lower bounds.
- Score: 2.750124853532831
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: A well-studied challenge that arises in the structure learning problem of
causal directed acyclic graphs (DAG) is that using observational data, one can
only learn the graph up to a "Markov equivalence class" (MEC). The remaining
undirected edges have to be oriented using interventions, which can be very
expensive to perform in applications. Thus, the problem of minimizing the
number of interventions needed to fully orient the MEC has received a lot of
recent attention, and is also the focus of this work. We prove two main
results. The first is a new universal lower bound on the number of atomic
interventions that any algorithm (whether active or passive) would need to
perform in order to orient a given MEC. Our second result shows that this bound
is, in fact, within a factor of two of the size of the smallest set of atomic
interventions that can orient the MEC. Our lower bound is provably better than
previously known lower bounds. The proof of our lower bound is based on the new
notion of CBSP orderings, which are topological orderings of DAGs without
v-structures and satisfy certain special properties. Further, using simulations
on synthetic graphs and by giving examples of special graph families, we show
that our bound is often significantly better.
Related papers
- Sample Efficient Bayesian Learning of Causal Graphs from Interventions [6.823521786512908]
This study considers a Bayesian approach for learning causal graphs with limited interventional samples.
We show theoretically that our proposed algorithm will return the true causal graph with high probability.
We present a case study showing how this algorithm could be modified to answer more general causal questions without learning the whole graph.
arXiv Detail & Related papers (2024-10-26T05:47:56Z) - Unraveling the Impact of Heterophilic Structures on Graph Positive-Unlabeled Learning [71.9954600831939]
Positive-Unlabeled (PU) learning is vital in many real-world scenarios, but its application to graph data remains under-explored.
We unveil that a critical challenge for PU learning on graph lies on the edge heterophily, which directly violates the irreducibility assumption for Class-Prior Estimation.
In response to this challenge, we introduce a new method, named Graph PU Learning with Label Propagation Loss (GPL)
arXiv Detail & Related papers (2024-05-30T10:30:44Z) - One-step Bipartite Graph Cut: A Normalized Formulation and Its
Application to Scalable Subspace Clustering [56.81492360414741]
We show how to enforce a one-step normalized cut for bipartite graphs, especially with linear-time complexity.
In this paper, we first characterize a novel one-step bipartite graph cut criterion with normalized constraints, and theoretically prove its equivalence to a trace problem.
We extend this cut criterion to a scalable subspace clustering approach, where adaptive anchor learning, bipartite graph learning, and one-step normalized bipartite graph partitioning are simultaneously modeled.
arXiv Detail & Related papers (2023-05-12T11:27:20Z) - Escaping limit cycles: Global convergence for constrained
nonconvex-nonconcave minimax problems [46.71914490521821]
This paper introduces a new extragradient-type algorithm for a class of non-nonconcave minimax problems.
The proposed algorithm is applicable to constrained and regularized problems.
arXiv Detail & Related papers (2023-02-20T08:37:08Z) - OrthoReg: Improving Graph-regularized MLPs via Orthogonality
Regularization [66.30021126251725]
Graph Neural Networks (GNNs) are currently dominating in modeling graphstructure data.
Graph-regularized networks (GR-MLPs) implicitly inject the graph structure information into model weights, while their performance can hardly match that of GNNs in most tasks.
We show that GR-MLPs suffer from dimensional collapse, a phenomenon in which the largest a few eigenvalues dominate the embedding space.
We propose OrthoReg, a novel GR-MLP model to mitigate the dimensional collapse issue.
arXiv Detail & Related papers (2023-01-31T21:20:48Z) - Subset verification and search algorithms for causal DAGs [13.108039226223793]
We study the problem of identifying the smallest set of interventions required to learn the causal relationships between a subset of edges (target edges)
We prove several interesting structural properties of interventional causal graphs that we believe have applications beyond the subset verification/search problems studied here.
arXiv Detail & Related papers (2023-01-09T06:25:44Z) - BCD Nets: Scalable Variational Approaches for Bayesian Causal Discovery [97.79015388276483]
A structural equation model (SEM) is an effective framework to reason over causal relationships represented via a directed acyclic graph (DAG)
Recent advances enabled effective maximum-likelihood point estimation of DAGs from observational data.
We propose BCD Nets, a variational framework for estimating a distribution over DAGs characterizing a linear-Gaussian SEM.
arXiv Detail & Related papers (2021-12-06T03:35:21Z) - Intervention Efficient Algorithm for Two-Stage Causal MDPs [15.838256272508357]
We study Markov Decision Processes (MDP) wherein states correspond to causal graphs that generate rewards.
In this setup, the learner's goal is to identify atomic interventions that lead to high rewards by intervening on variables at each state.
Generalizing the recent causal-bandit framework, the current work develops (simple) regret minimization guarantees for two-stage causal MDPs.
arXiv Detail & Related papers (2021-11-01T12:22:37Z) - Direction Matters: On the Implicit Bias of Stochastic Gradient Descent
with Moderate Learning Rate [105.62979485062756]
This paper attempts to characterize the particular regularization effect of SGD in the moderate learning rate regime.
We show that SGD converges along the large eigenvalue directions of the data matrix, while GD goes after the small eigenvalue directions.
arXiv Detail & Related papers (2020-11-04T21:07:52Z) - Active Structure Learning of Causal DAGs via Directed Clique Tree [28.637447727749]
We develop a universal lower bound for single-node interventions that establishes that the largest clique is always a fundamental impediment to structure learning.
Specifically, we present a decomposition of a DAG into independently orientable components through directed clique trees.
We also present a two-phase intervention design algorithm that matches the optimal number of interventions up to a multiplicative logarithmic factor in the number of maximal cliques.
arXiv Detail & Related papers (2020-11-01T23:11:17Z) - CLUB: A Contrastive Log-ratio Upper Bound of Mutual Information [105.73798100327667]
We propose a novel Contrastive Log-ratio Upper Bound (CLUB) of mutual information.
We provide a theoretical analysis of the properties of CLUB and its variational approximation.
Based on this upper bound, we introduce a MI minimization training scheme and further accelerate it with a negative sampling strategy.
arXiv Detail & Related papers (2020-06-22T05:36:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.