Causal Estimation of Exposure Shifts with Neural Networks: Evaluating
the Health Benefits of Stricter Air Quality Standards in the US
- URL: http://arxiv.org/abs/2302.02560v3
- Date: Wed, 6 Dec 2023 18:55:43 GMT
- Title: Causal Estimation of Exposure Shifts with Neural Networks: Evaluating
the Health Benefits of Stricter Air Quality Standards in the US
- Authors: Mauricio Tec, Oladimeji Mudele, Kevin Josey, Francesca Dominici
- Abstract summary: We develop a neural network method and its theoretical underpinnings to estimate shift-response functions.
We apply our method to data consisting of 68 million individuals and 27 million deaths across the U.S.
Our goal is to estimate, for the first time, the reduction in deaths that would result from this anticipated revision using causal methods.
- Score: 3.1952340441132474
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In policy research, one of the most critical analytic tasks is to estimate
the causal effect of a policy-relevant shift to the distribution of a
continuous exposure/treatment on an outcome of interest. We call this problem
shift-response function (SRF) estimation. Existing neural network methods
involving robust causal-effect estimators lack theoretical guarantees and
practical implementations for SRF estimation. Motivated by a key
policy-relevant question in public health, we develop a neural network method
and its theoretical underpinnings to estimate SRFs with robustness and
efficiency guarantees. We then apply our method to data consisting of 68
million individuals and 27 million deaths across the U.S. to estimate the
causal effect from revising the US National Ambient Air Quality Standards
(NAAQS) for PM 2.5 from 12 $\mu g/m^3$ to 9 $\mu g/m^3$. This change has been
recently proposed by the US Environmental Protection Agency (EPA). Our goal is
to estimate, for the first time, the reduction in deaths that would result from
this anticipated revision using causal methods for SRFs. Our proposed method,
called {T}argeted {R}egularization for {E}xposure {S}hifts with Neural
{Net}works (TRESNET), contributes to the neural network literature for causal
inference in two ways: first, it proposes a targeted regularization loss with
theoretical properties that ensure double robustness and achieves asymptotic
efficiency specific for SRF estimation; second, it enables loss functions from
the exponential family of distributions to accommodate non-continuous outcome
distributions (such as hospitalization or mortality counts). We complement our
application with benchmark experiments that demonstrate TRESNET's broad
applicability and competitiveness.
Related papers
- Extended Fiducial Inference for Individual Treatment Effects via Deep Neural Networks [7.916654052803723]
This work introduces the Double Neural Network (Double-NN) method to address the problem of individual treatment effect estimation.<n>Deep neural networks are used to model the treatment and control effect functions, while an additional neural network is employed to estimate their parameters.<n> Numerical results highlight the superior performance of the proposed Double-NN method compared to the conformal quantile regression (CQR) method in individual treatment effect estimation.
arXiv Detail & Related papers (2025-05-04T05:40:45Z) - Regularization via f-Divergence: An Application to Multi-Oxide Spectroscopic Analysis [0.9565934024763958]
We propose a novel regularization method based on f-divergence to constrain the distributional discrepancy between predictions and noisy targets.
We conduct experiments using spectra collected in a Mars-like environment by the remote-sensing instruments aboard the Curiosity and Perseverance rovers.
arXiv Detail & Related papers (2025-02-06T03:37:35Z) - Doubly Robust Causal Effect Estimation under Networked Interference via Targeted Learning [24.63284452991301]
We propose a doubly robust causal effect estimator under networked interference.
Specifically, we generalize the targeted learning technique into the networked interference setting.
We devise an end-to-end causal effect estimator by transforming the identified theoretical condition into a targeted loss.
arXiv Detail & Related papers (2024-05-06T10:49:51Z) - TNDDR: Efficient and doubly robust estimation of COVID-19 vaccine
effectiveness under the test-negative design [0.46426477564038643]
The test-negative design (TND) is susceptible to selection bias due to outcome-dependent sampling.
We propose a one-step doubly robust and locally efficient estimator called TNDDR (TND doubly robust)
We apply it to estimate COVID-19 VE in an administrative dataset of community-dwelling older people (aged $geq 60$y) in the province of Qu'ebec, Canada.
arXiv Detail & Related papers (2023-10-06T20:40:59Z) - Understanding, Predicting and Better Resolving Q-Value Divergence in
Offline-RL [86.0987896274354]
We first identify a fundamental pattern, self-excitation, as the primary cause of Q-value estimation divergence in offline RL.
We then propose a novel Self-Excite Eigenvalue Measure (SEEM) metric to measure the evolving property of Q-network at training.
For the first time, our theory can reliably decide whether the training will diverge at an early stage.
arXiv Detail & Related papers (2023-10-06T17:57:44Z) - Neural Importance Sampling for Rapid and Reliable Gravitational-Wave
Inference [59.040209568168436]
We first generate a rapid proposal for the Bayesian posterior using neural networks, and then attach importance weights based on the underlying likelihood and prior.
This provides (1) a corrected posterior free from network inaccuracies, (2) a performance diagnostic (the sample efficiency) for assessing the proposal and identifying failure cases, and (3) an unbiased estimate of the Bayesian evidence.
We carry out a large study analyzing 42 binary black hole mergers observed by LIGO and Virgo with the SEOBNRv4PHM and IMRPhenomHMXP waveform models.
arXiv Detail & Related papers (2022-10-11T18:00:02Z) - How Tempering Fixes Data Augmentation in Bayesian Neural Networks [22.188535244056016]
We show that tempering implicitly reduces the misspecification arising from modeling augmentations as i.i.d. data.
The temperature mimics the role of the effective sample size, reflecting the gain in information provided by the augmentations.
arXiv Detail & Related papers (2022-05-27T11:06:56Z) - Estimating the Uncertainty of Neural Network Forecasts for Influenza
Prevalence Using Web Search Activity [3.632189127068905]
Influenza is an infectious disease with the potential to become a pandemic.
Forecasting its prevalence is an important undertaking for planning an effective response.
arXiv Detail & Related papers (2021-05-26T09:45:23Z) - Learning Calibrated Uncertainties for Domain Shift: A Distributionally
Robust Learning Approach [150.8920602230832]
We propose a framework for learning calibrated uncertainties under domain shifts.
In particular, the density ratio estimation reflects the closeness of a target (test) sample to the source (training) distribution.
We show that our proposed method generates calibrated uncertainties that benefit downstream tasks.
arXiv Detail & Related papers (2020-10-08T02:10:54Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z) - Scaling Equilibrium Propagation to Deep ConvNets by Drastically Reducing
its Gradient Estimator Bias [65.13042449121411]
In practice, training a network with the gradient estimates provided by EP does not scale to visual tasks harder than MNIST.
We show that a bias in the gradient estimate of EP, inherent in the use of finite nudging, is responsible for this phenomenon.
We apply these techniques to train an architecture with asymmetric forward and backward connections, yielding a 13.2% test error.
arXiv Detail & Related papers (2020-06-06T09:36:07Z) - Neural Networks and Value at Risk [59.85784504799224]
We perform Monte-Carlo simulations of asset returns for Value at Risk threshold estimation.
Using equity markets and long term bonds as test assets, we investigate neural networks.
We find our networks when fed with substantially less data to perform significantly worse.
arXiv Detail & Related papers (2020-05-04T17:41:59Z) - The Counterfactual $\chi$-GAN [20.42556178617068]
Causal inference often relies on the counterfactual framework, which requires that treatment assignment is independent of the outcome.
This work proposes a generative adversarial network (GAN)-based model called the Counterfactual $chi$-GAN (cGAN)
arXiv Detail & Related papers (2020-01-09T17:23:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.