Meta-learning for heterogeneous treatment effect estimation with
closed-form solvers
- URL: http://arxiv.org/abs/2305.11353v1
- Date: Fri, 19 May 2023 00:07:38 GMT
- Title: Meta-learning for heterogeneous treatment effect estimation with
closed-form solvers
- Authors: Tomoharu Iwata, Yoichi Chikahara
- Abstract summary: This article proposes a meta-learning method for estimating the conditional average treatment effect (CATE) from a few observational data.
The proposed method learns how to estimate CATEs from multiple tasks and uses the knowledge for unseen tasks.
- Score: 30.343569752920754
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This article proposes a meta-learning method for estimating the conditional
average treatment effect (CATE) from a few observational data. The proposed
method learns how to estimate CATEs from multiple tasks and uses the knowledge
for unseen tasks. In the proposed method, based on the meta-learner framework,
we decompose the CATE estimation problem into sub-problems. For each
sub-problem, we formulate our estimation models using neural networks with
task-shared and task-specific parameters. With our formulation, we can obtain
optimal task-specific parameters in a closed form that are differentiable with
respect to task-shared parameters, making it possible to perform effective
meta-learning. The task-shared parameters are trained such that the expected
CATE estimation performance in few-shot settings is improved by minimizing the
difference between a CATE estimated with a large amount of data and one
estimated with just a few data. Our experimental results demonstrate that our
method outperforms the existing meta-learning approaches and CATE estimation
methods.
Related papers
- Estimating Conditional Average Treatment Effects via Sufficient Representation Learning [31.822980052107496]
This paper proposes a novel neural network approach named textbfCrossNet to learn a sufficient representation for the features, based on which we then estimate the conditional average treatment effects (CATE)
Numerical simulations and empirical results demonstrate that our method outperforms the competitive approaches.
arXiv Detail & Related papers (2024-08-30T07:23:59Z) - Robust CATE Estimation Using Novel Ensemble Methods [0.8246494848934447]
estimation of Conditional Average Treatment Effects (CATE) is crucial for understanding the heterogeneity of treatment effects in clinical trials.
We evaluate the performance of common methods, including causal forests and various meta-learners, across a diverse set of scenarios.
We propose two new ensemble methods that integrate multiple estimators to enhance prediction stability and performance.
arXiv Detail & Related papers (2024-07-04T07:23:02Z) - MAP: Low-compute Model Merging with Amortized Pareto Fronts via Quadratic Approximation [80.47072100963017]
We introduce a novel and low-compute algorithm, Model Merging with Amortized Pareto Front (MAP)
MAP efficiently identifies a set of scaling coefficients for merging multiple models, reflecting the trade-offs involved.
We also introduce Bayesian MAP for scenarios with a relatively low number of tasks and Nested MAP for situations with a high number of tasks, further reducing the computational cost of evaluation.
arXiv Detail & Related papers (2024-06-11T17:55:25Z) - Estimation of conditional average treatment effects on distributed confidential data [6.798254568821052]
conditional average treatment effects (CATEs) can be estimated with high accuracy if distributed data across multiple parties can be centralized.
It is difficult to aggregate such data owing to confidential or privacy concerns.
We proposed data collaboration double machine learning, a method that can estimate CATE models from privacy-preserving fusion data constructed from distributed data.
arXiv Detail & Related papers (2024-02-05T02:17:21Z) - Task-Distributionally Robust Data-Free Meta-Learning [99.56612787882334]
Data-Free Meta-Learning (DFML) aims to efficiently learn new tasks by leveraging multiple pre-trained models without requiring their original training data.
For the first time, we reveal two major challenges hindering their practical deployments: Task-Distribution Shift ( TDS) and Task-Distribution Corruption (TDC)
arXiv Detail & Related papers (2023-11-23T15:46:54Z) - B-Learner: Quasi-Oracle Bounds on Heterogeneous Causal Effects Under
Hidden Confounding [51.74479522965712]
We propose a meta-learner called the B-Learner, which can efficiently learn sharp bounds on the CATE function under limits on hidden confounding.
We prove its estimates are valid, sharp, efficient, and have a quasi-oracle property with respect to the constituent estimators under more general conditions than existing methods.
arXiv Detail & Related papers (2023-04-20T18:07:19Z) - Task Difficulty Aware Parameter Allocation & Regularization for Lifelong
Learning [20.177260510548535]
We propose the Allocation & Regularization (PAR), which adaptively select an appropriate strategy for each task from parameter allocation and regularization based on its learning difficulty.
Our method is scalable and significantly reduces the model's redundancy while improving the model's performance.
arXiv Detail & Related papers (2023-04-11T15:38:21Z) - Time Associated Meta Learning for Clinical Prediction [78.99422473394029]
We propose a novel time associated meta learning (TAML) method to make effective predictions at multiple future time points.
To address the sparsity problem after task splitting, TAML employs a temporal information sharing strategy to augment the number of positive samples.
We demonstrate the effectiveness of TAML on multiple clinical datasets, where it consistently outperforms a range of strong baselines.
arXiv Detail & Related papers (2023-03-05T03:54:54Z) - MARS: Meta-Learning as Score Matching in the Function Space [79.73213540203389]
We present a novel approach to extracting inductive biases from a set of related datasets.
We use functional Bayesian neural network inference, which views the prior as a process and performs inference in the function space.
Our approach can seamlessly acquire and represent complex prior knowledge by metalearning the score function of the data-generating process.
arXiv Detail & Related papers (2022-10-24T15:14:26Z) - Probabilistic task modelling for meta-learning [44.072592379328036]
We propose a generative probabilistic model for collections of tasks used in meta-learning.
The proposed model combines variational auto-encoding and latent Dirichlet allocation to model each task as a mixture of Gaussian distribution in an embedding space.
arXiv Detail & Related papers (2021-06-09T04:34:12Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.