Learning Optimal Distributionally Robust Individualized Treatment Rules
- URL: http://arxiv.org/abs/2006.15121v1
- Date: Fri, 26 Jun 2020 17:24:25 GMT
- Title: Learning Optimal Distributionally Robust Individualized Treatment Rules
- Authors: Weibin Mo, Zhengling Qi, Yufeng Liu
- Abstract summary: Policy makers best individualized treatment rule (ITR) that maximizes the expected outcome, known as the value function.
Existing methods assume that the training and testing distributions are the same.
We propose a novel distributionally robust ITR framework that maximizes the worst-case value function.
- Score: 3.872376323711463
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent development in the data-driven decision science has seen great
advances in individualized decision making. Given data with individual
covariates, treatment assignments and outcomes, policy makers best
individualized treatment rule (ITR) that maximizes the expected outcome, known
as the value function. Many existing methods assume that the training and
testing distributions are the same. However, the estimated optimal ITR may have
poor generalizability when the training and testing distributions are not
identical. In this paper, we consider the problem of finding an optimal ITR
from a restricted ITR class where there is some unknown covariate changes
between the training and testing distributions. We propose a novel
distributionally robust ITR (DR-ITR) framework that maximizes the worst-case
value function across the values under a set of underlying distributions that
are "close" to the training distribution. The resulting DR-ITR can guarantee
the performance among all such distributions reasonably well. We further
propose a calibrating procedure that tunes the DR-ITR adaptively to a small
amount of calibration data from a target population. In this way, the
calibrated DR-ITR can be shown to enjoy better generalizability than the
standard ITR based on our numerical studies.
Related papers
- DRAUC: An Instance-wise Distributionally Robust AUC Optimization
Framework [133.26230331320963]
Area Under the ROC Curve (AUC) is a widely employed metric in long-tailed classification scenarios.
We propose an instance-wise surrogate loss of Distributionally Robust AUC (DRAUC) and build our optimization framework on top of it.
arXiv Detail & Related papers (2023-11-06T12:15:57Z) - Safe Deployment for Counterfactual Learning to Rank with Exposure-Based
Risk Minimization [63.93275508300137]
We introduce a novel risk-aware Counterfactual Learning To Rank method with theoretical guarantees for safe deployment.
Our experimental results demonstrate the efficacy of our proposed method, which is effective at avoiding initial periods of bad performance when little data is available.
arXiv Detail & Related papers (2023-04-26T15:54:23Z) - Policy Evaluation in Distributional LQR [70.63903506291383]
We provide a closed-form expression of the distribution of the random return.
We show that this distribution can be approximated by a finite number of random variables.
Using the approximate return distribution, we propose a zeroth-order policy gradient algorithm for risk-averse LQR.
arXiv Detail & Related papers (2023-03-23T20:27:40Z) - Efficient and robust transfer learning of optimal individualized
treatment regimes with right-censored survival data [7.308241944759317]
An individualized treatment regime (ITR) is a decision rule that assigns treatments based on patients' characteristics.
We propose a doubly robust estimator of the value function, and the optimal ITR is learned by maximizing the value function within a pre-specified class of ITRs.
We evaluate the empirical performance of the proposed method by simulation studies and a real data application of sodium bicarbonate therapy for patients with severe metabolic acidaemia.
arXiv Detail & Related papers (2023-01-13T11:47:10Z) - Normality-Guided Distributional Reinforcement Learning for Continuous
Control [16.324313304691426]
Learning a predictive model of the mean return, or value function, plays a critical role in many reinforcement learning algorithms.
We study the value distribution in several continuous control tasks and find that the learned value distribution is empirical quite close to normal.
We propose a policy update strategy based on the correctness as measured by structural characteristics of the value distribution not present in the standard value function.
arXiv Detail & Related papers (2022-08-28T02:52:10Z) - Distributionally Robust Models with Parametric Likelihood Ratios [123.05074253513935]
Three simple ideas allow us to train models with DRO using a broader class of parametric likelihood ratios.
We find that models trained with the resulting parametric adversaries are consistently more robust to subpopulation shifts when compared to other DRO approaches.
arXiv Detail & Related papers (2022-04-13T12:43:12Z) - Minimax Regret Optimization for Robust Machine Learning under
Distribution Shift [38.30154154957721]
We consider learning scenarios where the learned model is evaluated under an unknown test distribution.
We show that the DRO formulation does not guarantee uniformly small regret under distribution shift.
We propose an alternative method called Minimax Regret Optimization (MRO)
arXiv Detail & Related papers (2022-02-11T04:17:22Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Distributionally Robust Multi-Output Regression Ranking [3.9318191265352196]
We introduce a new listwise listwise learning-to-rank model called Distributionally Robust Multi-output Regression Ranking (DRMRR)
DRMRR uses a Distributionally Robust Optimization framework to minimize a multi-output loss function under the most adverse distributions in the neighborhood of the empirical data distribution.
Our experiments were conducted on two real-world applications, medical document retrieval, and drug response prediction.
arXiv Detail & Related papers (2021-09-27T05:19:27Z) - Learning Calibrated Uncertainties for Domain Shift: A Distributionally
Robust Learning Approach [150.8920602230832]
We propose a framework for learning calibrated uncertainties under domain shifts.
In particular, the density ratio estimation reflects the closeness of a target (test) sample to the source (training) distribution.
We show that our proposed method generates calibrated uncertainties that benefit downstream tasks.
arXiv Detail & Related papers (2020-10-08T02:10:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.