Individual Treatment Effect Estimation Through Controlled Neural Network
Training in Two Stages
- URL: http://arxiv.org/abs/2201.08559v1
- Date: Fri, 21 Jan 2022 06:34:52 GMT
- Title: Individual Treatment Effect Estimation Through Controlled Neural Network
Training in Two Stages
- Authors: Naveen Nair, Karthik S. Gurumoorthy, Dinesh Mandalapu
- Abstract summary: We develop a Causal-Deep Neural Network model trained in two stages to infer causal impact estimates at an individual unit level.
We observe that CDNN is highly competitive and often yields the most accurate individual treatment effect estimates.
- Score: 0.757024681220677
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We develop a Causal-Deep Neural Network (CDNN) model trained in two stages to
infer causal impact estimates at an individual unit level. Using only the
pre-treatment features in stage 1 in the absence of any treatment information,
we learn an encoding for the covariates that best represents the outcome. In
the $2^{nd}$ stage we further seek to predict the unexplained outcome from
stage 1, by introducing the treatment indicator variables alongside the encoded
covariates. We prove that even without explicitly computing the treatment
residual, our method still satisfies the desirable local Neyman orthogonality,
making it robust to small perturbations in the nuisance parameters.
Furthermore, by establishing connections with the representation learning
approaches, we create a framework from which multiple variants of our algorithm
can be derived. We perform initial experiments on the publicly available data
sets to compare these variants and get guidance in selecting the best variant
of our CDNN method. On evaluating CDNN against the state-of-the-art approaches
on three benchmarking datasets, we observe that CDNN is highly competitive and
often yields the most accurate individual treatment effect estimates. We
highlight the strong merits of CDNN in terms of its extensibility to multiple
use cases.
Related papers
- Causal inference through multi-stage learning and doubly robust deep neural networks [10.021381302215062]
Deep neural networks (DNNs) have demonstrated remarkable empirical performance in large-scale supervised learning problems.
This study delves into the application of DNNs across a wide spectrum of intricate causal inference tasks.
arXiv Detail & Related papers (2024-07-11T14:47:44Z) - Continuous Treatment Effect Estimation Using Gradient Interpolation and
Kernel Smoothing [43.259723628010896]
We advocate the direct approach of augmenting training individuals with independently sampled treatments and inferred counterfactual outcomes.
We evaluate our method on five benchmarks and show that our method outperforms six state-of-the-art methods on the counterfactual estimation error.
arXiv Detail & Related papers (2024-01-27T15:52:58Z) - Estimation of individual causal effects in network setup for multiple
treatments [4.53340898566495]
We study the problem of estimation of Individual Treatment Effects (ITE) in the context of multiple treatments and observational data.
We employ Graph Convolutional Networks (GCN) to learn a shared representation of the confounders.
Our approach utilizes separate neural networks to infer potential outcomes for each treatment.
arXiv Detail & Related papers (2023-12-18T06:07:45Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - Bridging Precision and Confidence: A Train-Time Loss for Calibrating
Object Detection [58.789823426981044]
We propose a novel auxiliary loss formulation that aims to align the class confidence of bounding boxes with the accurateness of predictions.
Our results reveal that our train-time loss surpasses strong calibration baselines in reducing calibration error for both in and out-domain scenarios.
arXiv Detail & Related papers (2023-03-25T08:56:21Z) - Rethinking Semi-Supervised Medical Image Segmentation: A
Variance-Reduction Perspective [51.70661197256033]
We propose ARCO, a semi-supervised contrastive learning framework with stratified group theory for medical image segmentation.
We first propose building ARCO through the concept of variance-reduced estimation and show that certain variance-reduction techniques are particularly beneficial in pixel/voxel-level segmentation tasks.
We experimentally validate our approaches on eight benchmarks, i.e., five 2D/3D medical and three semantic segmentation datasets, with different label settings.
arXiv Detail & Related papers (2023-02-03T13:50:25Z) - Neuron Coverage-Guided Domain Generalization [37.77033512313927]
This paper focuses on the domain generalization task where domain knowledge is unavailable, and even worse, only samples from a single domain can be utilized during training.
Our motivation originates from the recent progresses in deep neural network (DNN) testing, which has shown that maximizing neuron coverage of DNN can help to explore possible defects of DNN.
arXiv Detail & Related papers (2021-02-27T14:26:53Z) - Targeted Attack against Deep Neural Networks via Flipping Limited Weight
Bits [55.740716446995805]
We study a novel attack paradigm, which modifies model parameters in the deployment stage for malicious purposes.
Our goal is to misclassify a specific sample into a target class without any sample modification.
By utilizing the latest technique in integer programming, we equivalently reformulate this BIP problem as a continuous optimization problem.
arXiv Detail & Related papers (2021-02-21T03:13:27Z) - S2-BNN: Bridging the Gap Between Self-Supervised Real and 1-bit Neural
Networks via Guided Distribution Calibration [74.5509794733707]
We present a novel guided learning paradigm from real-valued to distill binary networks on the final prediction distribution.
Our proposed method can boost the simple contrastive learning baseline by an absolute gain of 5.515% on BNNs.
Our method achieves substantial improvement over the simple contrastive learning baseline, and is even comparable to many mainstream supervised BNN methods.
arXiv Detail & Related papers (2021-02-17T18:59:28Z) - An Uncertainty-Driven GCN Refinement Strategy for Organ Segmentation [53.425900196763756]
We propose a segmentation refinement method based on uncertainty analysis and graph convolutional networks.
We employ the uncertainty levels of the convolutional network in a particular input volume to formulate a semi-supervised graph learning problem.
We show that our method outperforms the state-of-the-art CRF refinement method by improving the dice score by 1% for the pancreas and 2% for spleen.
arXiv Detail & Related papers (2020-12-06T18:55:07Z) - Revisiting One-vs-All Classifiers for Predictive Uncertainty and
Out-of-Distribution Detection in Neural Networks [22.34227625637843]
We investigate how the parametrization of the probabilities in discriminative classifiers affects the uncertainty estimates.
We show that one-vs-all formulations can improve calibration on image classification tasks.
arXiv Detail & Related papers (2020-07-10T01:55:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.