Balance Regularized Neural Network Models for Causal Effect Estimation
- URL: http://arxiv.org/abs/2011.11199v1
- Date: Mon, 23 Nov 2020 04:03:55 GMT
- Title: Balance Regularized Neural Network Models for Causal Effect Estimation
- Authors: Mehrdad Farajtabar, Andrew Lee, Yuanjian Feng, Vishal Gupta, Peter
Dolan, Harish Chandran, Martin Szummer
- Abstract summary: We advocate balance regularization of multi-head neural network architectures.
We further regularize the model by encouraging it to predict control outcomes for individuals in the treatment group that are similar to control outcomes in the control group.
- Score: 16.8658322310041
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Estimating individual and average treatment effects from observational data
is an important problem in many domains such as healthcare and e-commerce. In
this paper, we advocate balance regularization of multi-head neural network
architectures. Our work is motivated by representation learning techniques to
reduce differences between treated and untreated distributions that potentially
arise due to confounding factors. We further regularize the model by
encouraging it to predict control outcomes for individuals in the treatment
group that are similar to control outcomes in the control group. We empirically
study the bias-variance trade-off between different weightings of the
regularizers, as well as between inductive and transductive inference.
Related papers
- Deep Disentangled Representation Network for Treatment Effect Estimation [17.787544758117377]
Estimating individual-level treatment effect from observational data is a fundamental problem in causal inference.<n>We propose a novel treatment effect estimation algorithm that incorporates a mixture of experts with multi-head attention.<n>We conduct extensive experiments on both public semi-synthetic and real-world production datasets.
arXiv Detail & Related papers (2025-07-09T08:29:37Z) - Conformal Diffusion Models for Individual Treatment Effect Estimation and Inference [6.406853903837333]
Individual treatment effect offers the most granular measure of treatment effect on an individual level.
We propose a novel conformal diffusion model-based approach that addresses those intricate challenges.
arXiv Detail & Related papers (2024-08-02T21:35:08Z) - Model-Based Inference and Experimental Design for Interference Using Partial Network Data [4.76518127830168]
We present a framework for the estimation and inference of treatment effect adjustments using partial network data.
We illustrate procedures to assign treatments using only partial network data.
We validate our approach using simulated experiments on observed graphs with applications to information diffusion in India and Malawi.
arXiv Detail & Related papers (2024-06-17T17:27:18Z) - De-confounding Representation Learning for Counterfactual Inference on
Continuous Treatment via Generative Adversarial Network [5.465397606401007]
Counterfactual inference for continuous rather than binary treatment variables is more common in real-world causal inference tasks.
We propose a de-confounding representation learning (DRL) framework for counterfactual outcome estimation of continuous treatment.
We show that the DRL model performs superiorly in learning de-confounding representations and outperforms state-of-the-art counterfactual inference models for continuous treatment variables.
arXiv Detail & Related papers (2023-07-24T08:56:25Z) - Analyzing the Effects of Handling Data Imbalance on Learned Features
from Medical Images by Looking Into the Models [50.537859423741644]
Training a model on an imbalanced dataset can introduce unique challenges to the learning problem.
We look deeper into the internal units of neural networks to observe how handling data imbalance affects the learned features.
arXiv Detail & Related papers (2022-04-04T09:38:38Z) - Covariate-Balancing-Aware Interpretable Deep Learning models for
Treatment Effect Estimation [15.465045049754336]
We propose an upper bound for the bias of average treatment estimation under the strong ignorability assumption.
We implement this upper bound as an objective function being minimized by leveraging a novel additive neural network architecture.
The proposed method is illustrated by re-examining the benchmark datasets for causal inference, and it outperforms the state-of-art.
arXiv Detail & Related papers (2022-03-07T07:42:40Z) - On the Robustness of Pretraining and Self-Supervision for a Deep
Learning-based Analysis of Diabetic Retinopathy [70.71457102672545]
We compare the impact of different training procedures for diabetic retinopathy grading.
We investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions.
Our results indicate that models from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions.
arXiv Detail & Related papers (2021-06-25T08:32:45Z) - Efficient Causal Inference from Combined Observational and
Interventional Data through Causal Reductions [68.6505592770171]
Unobserved confounding is one of the main challenges when estimating causal effects.
We propose a novel causal reduction method that replaces an arbitrary number of possibly high-dimensional latent confounders.
We propose a learning algorithm to estimate the parameterized reduced model jointly from observational and interventional data.
arXiv Detail & Related papers (2021-03-08T14:29:07Z) - Learning Decomposed Representation for Counterfactual Inference [53.36586760485262]
The fundamental problem in treatment effect estimation from observational data is confounder identification and balancing.
Most of the previous methods realized confounder balancing by treating all observed pre-treatment variables as confounders, ignoring further identifying confounders and non-confounders.
We propose a synergistic learning framework to 1) identify confounders by learning representations of both confounders and non-confounders, 2) balance confounder with sample re-weighting technique, and simultaneously 3) estimate the treatment effect in observational studies via counterfactual inference.
arXiv Detail & Related papers (2020-06-12T09:50:42Z) - Estimating the Effects of Continuous-valued Interventions using
Generative Adversarial Networks [103.14809802212535]
We build on the generative adversarial networks (GANs) framework to address the problem of estimating the effect of continuous-valued interventions.
Our model, SCIGAN, is flexible and capable of simultaneously estimating counterfactual outcomes for several different continuous interventions.
To address the challenges presented by shifting to continuous interventions, we propose a novel architecture for our discriminator.
arXiv Detail & Related papers (2020-02-27T18:46:21Z) - Generalization Bounds and Representation Learning for Estimation of
Potential Outcomes and Causal Effects [61.03579766573421]
We study estimation of individual-level causal effects, such as a single patient's response to alternative medication.
We devise representation learning algorithms that minimize our bound, by regularizing the representation's induced treatment group distance.
We extend these algorithms to simultaneously learn a weighted representation to further reduce treatment group distances.
arXiv Detail & Related papers (2020-01-21T10:16:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.