Jump Interval-Learning for Individualized Decision Making
- URL: http://arxiv.org/abs/2111.08885v1
- Date: Wed, 17 Nov 2021 03:29:59 GMT
- Title: Jump Interval-Learning for Individualized Decision Making
- Authors: Hengrui Cai, Chengchun Shi, Rui Song, Wenbin Lu
- Abstract summary: We propose a jump interval-learning to develop an individualized interval-valued decision rule (I2DR)
Unlike IDRs that recommend a single treatment, the proposed I2DR yields an interval of treatment options for each individual.
- Score: 21.891586204541877
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An individualized decision rule (IDR) is a decision function that assigns
each individual a given treatment based on his/her observed characteristics.
Most of the existing works in the literature consider settings with binary or
finitely many treatment options. In this paper, we focus on the continuous
treatment setting and propose a jump interval-learning to develop an
individualized interval-valued decision rule (I2DR) that maximizes the expected
outcome. Unlike IDRs that recommend a single treatment, the proposed I2DR
yields an interval of treatment options for each individual, making it more
flexible to implement in practice. To derive an optimal I2DR, our jump
interval-learning method estimates the conditional mean of the outcome given
the treatment and the covariates via jump penalized regression, and derives the
corresponding optimal I2DR based on the estimated outcome regression function.
The regressor is allowed to be either linear for clear interpretation or deep
neural network to model complex treatment-covariates interactions. To implement
jump interval-learning, we develop a searching algorithm based on dynamic
programming that efficiently computes the outcome regression function.
Statistical properties of the resulting I2DR are established when the outcome
regression function is either a piecewise or continuous function over the
treatment space. We further develop a procedure to infer the mean outcome under
the (estimated) optimal policy. Extensive simulations and a real data
application to a warfarin study are conducted to demonstrate the empirical
validity of the proposed I2DR.
Related papers
- Robust Learning for Optimal Dynamic Treatment Regimes with Observational Data [0.0]
We study the statistical learning of optimal dynamic treatment regimes (DTRs) that guide the optimal treatment assignment for each individual at each stage based on the individual's evolving history.
arXiv Detail & Related papers (2024-03-30T02:33:39Z) - Doubly Robust Proximal Causal Learning for Continuous Treatments [56.05592840537398]
We propose a kernel-based doubly robust causal learning estimator for continuous treatments.
We show that its oracle form is a consistent approximation of the influence function.
We then provide a comprehensive convergence analysis in terms of the mean square error.
arXiv Detail & Related papers (2023-09-22T12:18:53Z) - Efficient and robust transfer learning of optimal individualized
treatment regimes with right-censored survival data [7.308241944759317]
An individualized treatment regime (ITR) is a decision rule that assigns treatments based on patients' characteristics.
We propose a doubly robust estimator of the value function, and the optimal ITR is learned by maximizing the value function within a pre-specified class of ITRs.
We evaluate the empirical performance of the proposed method by simulation studies and a real data application of sodium bicarbonate therapy for patients with severe metabolic acidaemia.
arXiv Detail & Related papers (2023-01-13T11:47:10Z) - Nearly Minimax Optimal Reinforcement Learning for Linear Markov Decision
Processes [80.89852729380425]
We propose the first computationally efficient algorithm that achieves the nearly minimax optimal regret $tilde O(dsqrtH3K)$.
Our work provides a complete answer to optimal RL with linear MDPs, and the developed algorithm and theoretical tools may be of independent interest.
arXiv Detail & Related papers (2022-12-12T18:58:59Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Robust and Agnostic Learning of Conditional Distributional Treatment
Effects [62.44901952244514]
The conditional average treatment effect (CATE) is the best point prediction of individual causal effects.
In aggregate analyses, this is usually addressed by measuring distributional treatment effect (DTE)
We provide a new robust and model-agnostic methodology for learning the conditional DTE (CDTE) for a wide class of problems.
arXiv Detail & Related papers (2022-05-23T17:40:31Z) - Variance-Aware Off-Policy Evaluation with Linear Function Approximation [85.75516599931632]
We study the off-policy evaluation problem in reinforcement learning with linear function approximation.
We propose an algorithm, VA-OPE, which uses the estimated variance of the value function to reweight the Bellman residual in Fitted Q-Iteration.
arXiv Detail & Related papers (2021-06-22T17:58:46Z) - Machine Learning based optimization for interval uncertainty propagation
with application to vibro-acoustic models [0.0]
Two non-intrusive uncertainty propagation approaches are proposed for the performance analysis of engineering systems.
One approach builds iteratively two distinct training datasets for evaluating separately the upper and lower bounds of the response variable.
The other builds iteratively a single training dataset.
arXiv Detail & Related papers (2021-06-21T15:57:11Z) - Evaluating (weighted) dynamic treatment effects by double machine
learning [0.12891210250935145]
We consider evaluating the causal effects of dynamic treatments in a data-driven way under a selection-on-observables assumption.
We make use of so-called Neyman-orthogonal score functions, which imply the robustness of treatment effect estimation to moderate (local) misspecifications.
We demonstrate that the estimators are regularityally normal and $sqrtn$-consistent under specific conditions.
arXiv Detail & Related papers (2020-12-01T09:55:40Z) - Leveraging both Lesion Features and Procedural Bias in Neuroimaging: An
Dual-Task Split dynamics of inverse scale space [21.05070956384346]
The prediction and selection of lesion features are two important tasks in voxel-based neuroimage analysis.
In this paper, we propose that the features/voxels in neuroimage data are consist of three parts: lesion features, procedural bias, and null features.
To stably select lesion features and leverage procedural bias into prediction, we propose an iterative algorithm (termed GSplit LBI)
The validity and the benefit of our model can be shown by the improvement of prediction results and the interpretability of visualized procedural bias and lesion features.
arXiv Detail & Related papers (2020-07-17T03:41:48Z) - DTR Bandit: Learning to Make Response-Adaptive Decisions With Low Regret [59.81290762273153]
Dynamic treatment regimes (DTRs) are personalized, adaptive, multi-stage treatment plans that adapt treatment decisions to an individual's initial features and to intermediate outcomes and features at each subsequent stage.
We propose a novel algorithm that, by carefully balancing exploration and exploitation, is guaranteed to achieve rate-optimal regret when the transition and reward models are linear.
arXiv Detail & Related papers (2020-05-06T13:03:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.