H\"older Bounds for Sensitivity Analysis in Causal Reasoning
- URL: http://arxiv.org/abs/2107.04661v1
- Date: Fri, 9 Jul 2021 20:26:36 GMT
- Title: H\"older Bounds for Sensitivity Analysis in Causal Reasoning
- Authors: Serge Assaad, Shuxi Zeng, Henry Pfister, Fan Li, Lawrence Carin
- Abstract summary: We derive a set of bounds on the confounding bias |E[Y|T=t]-E[Y|do(T=t)]| based on the degree of unmeasured confounding.
These bounds are tight either when U is independent of T or when U is independent of Y given T.
- Score: 66.00472443147781
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We examine interval estimation of the effect of a treatment T on an outcome Y
given the existence of an unobserved confounder U. Using H\"older's inequality,
we derive a set of bounds on the confounding bias |E[Y|T=t]-E[Y|do(T=t)]| based
on the degree of unmeasured confounding (i.e., the strength of the connection
U->T, and the strength of U->Y). These bounds are tight either when U is
independent of T or when U is independent of Y given T (when there is no
unobserved confounding). We focus on a special case of this bound depending on
the total variation distance between the distributions p(U) and p(U|T=t), as
well as the maximum (over all possible values of U) deviation of the
conditional expected outcome E[Y|U=u,T=t] from the average expected outcome
E[Y|T=t]. We discuss possible calibration strategies for this bound to get
interval estimates for treatment effects, and experimentally validate the bound
using synthetic and semi-synthetic datasets.
Related papers
- On Statistical Rates of Conditional Diffusion Transformers: Approximation, Estimation and Minimax Optimality [15.889816082916722]
We show that both conditional DiTs and their latent variants lead to the minimax optimality of unconditional DiTs under identified settings.
Our findings establish statistical limits for conditional and unconditional DiTs, and offer practical guidance toward developing more efficient and accurate DiT models.
arXiv Detail & Related papers (2024-11-26T15:30:48Z) - A Unified Analysis for Finite Weight Averaging [50.75116992029417]
Averaging iterations of Gradient Descent (SGD) have achieved empirical success in training deep learning models, such as Weight Averaging (SWA), Exponential Moving Average (EMA), and LAtest Weight Averaging (LAWA)
In this paper, we generalize LAWA as Finite Weight Averaging (FWA) and explain their advantages compared to SGD from the perspective of optimization and generalization.
arXiv Detail & Related papers (2024-11-20T10:08:22Z) - Estimation of entropy-regularized optimal transport maps between
non-compactly supported measures [15.857723276537248]
This paper addresses the problem of estimating entropy-regularized optimal transport maps with squared-Euclidean cost between source and target measures that are subGaussian.
arXiv Detail & Related papers (2023-11-20T17:18:21Z) - TIC-TAC: A Framework for Improved Covariance Estimation in Deep Heteroscedastic Regression [109.69084997173196]
Deepscedastic regression involves jointly optimizing the mean and covariance of the predicted distribution using the negative log-likelihood.
Recent works show that this may result in sub-optimal convergence due to the challenges associated with covariance estimation.
We study two questions: (1) Does the predicted covariance truly capture the randomness of the predicted mean?
Our results show that not only does TIC accurately learn the covariance, it additionally facilitates an improved convergence of the negative log-likelihood.
arXiv Detail & Related papers (2023-10-29T09:54:03Z) - A High-dimensional Convergence Theorem for U-statistics with
Applications to Kernel-based Testing [3.469038201881982]
We prove a convergence theorem for U-statistics of degree two, where the data dimension $d$ is allowed to scale with sample size $n$.
We apply our theory to two popular kernel-based distribution tests, MMD and KSD, whose high-dimensional performance has been challenging to study.
arXiv Detail & Related papers (2023-02-11T12:49:46Z) - Robust computation of optimal transport by $\beta$-potential
regularization [79.24513412588745]
Optimal transport (OT) has become a widely used tool in the machine learning field to measure the discrepancy between probability distributions.
We propose regularizing OT with the beta-potential term associated with the so-called $beta$-divergence.
We experimentally demonstrate that the transport matrix computed with our algorithm helps estimate a probability distribution robustly even in the presence of outliers.
arXiv Detail & Related papers (2022-12-26T18:37:28Z) - A lower confidence sequence for the changing mean of non-negative right
heavy-tailed observations with bounded mean [9.289846887298854]
A confidence sequence produces an adapted sequence of sets for a predictable parameter sequence with a time-parametric coverage guarantee.
This work constructs a non-asymptotic lower CS for the running average conditional expectation whose slack converges to zero.
arXiv Detail & Related papers (2022-10-20T09:50:05Z) - Statistical Efficiency of Score Matching: The View from Isoperimetry [96.65637602827942]
We show a tight connection between statistical efficiency of score matching and the isoperimetric properties of the distribution being estimated.
We formalize these results both in the sample regime and in the finite regime.
arXiv Detail & Related papers (2022-10-03T06:09:01Z) - On Linear Stochastic Approximation: Fine-grained Polyak-Ruppert and
Non-Asymptotic Concentration [115.1954841020189]
We study the inequality and non-asymptotic properties of approximation procedures with Polyak-Ruppert averaging.
We prove a central limit theorem (CLT) for the averaged iterates with fixed step size and number of iterations going to infinity.
arXiv Detail & Related papers (2020-04-09T17:54:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.