Kernel-Based Distributed Q-Learning: A Scalable Reinforcement Learning
Approach for Dynamic Treatment Regimes
- URL: http://arxiv.org/abs/2302.10434v1
- Date: Tue, 21 Feb 2023 04:15:34 GMT
- Title: Kernel-Based Distributed Q-Learning: A Scalable Reinforcement Learning
Approach for Dynamic Treatment Regimes
- Authors: Di Wang, Yao Wang, Shaojie Tang, Shao-Bo Lin
- Abstract summary: We present a novel approach, a distributed Q-learning algorithm, for generating dynamic treatment regimes.
We show that our algorithm outperforms both traditional linear Q-learning and commonly used deep Q-learning in terms of both prediction accuracy and computation cost.
- Score: 37.06048335758881
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, large amounts of electronic health records (EHRs) concerning
chronic diseases, such as cancer, diabetes, and mental disease, have been
collected to facilitate medical diagnosis. Modeling the dynamic properties of
EHRs related to chronic diseases can be efficiently done using dynamic
treatment regimes (DTRs), which are a set of sequential decision rules. While
Reinforcement learning (RL) is a widely used method for creating DTRs, there is
ongoing research in developing RL algorithms that can effectively handle large
amounts of data. In this paper, we present a novel approach, a distributed
Q-learning algorithm, for generating DTRs. The novelties of our research are as
follows: 1) From a methodological perspective, we present a novel and scalable
approach for generating DTRs by combining distributed learning with Q-learning.
The proposed approach is specifically designed to handle large amounts of data
and effectively generate DTRs. 2) From a theoretical standpoint, we provide
generalization error bounds for the proposed distributed Q-learning algorithm,
which are derived within the framework of statistical learning theory. These
bounds quantify the relationships between sample size, prediction accuracy, and
computational burden, providing insights into the performance of the algorithm.
3) From an applied perspective, we demonstrate the effectiveness of our
proposed distributed Q-learning algorithm for DTRs by applying it to clinical
cancer treatments. The results show that our algorithm outperforms both
traditional linear Q-learning and commonly used deep Q-learning in terms of
both prediction accuracy and computation cost.
Related papers
- Fine-tuning -- a Transfer Learning approach [0.22344294014777952]
Missingness in Electronic Health Records (EHRs) is often hampered by the abundance of missing data in this valuable resource.
Existing deep imputation methods rely on end-to-end pipelines that incorporate both imputation and downstream analyses.
This paper explores the development of a modular, deep learning-based imputation and classification pipeline.
arXiv Detail & Related papers (2024-11-06T14:18:23Z) - DTR-Bench: An in silico Environment and Benchmark Platform for Reinforcement Learning Based Dynamic Treatment Regime [18.443316087890324]
Reinforcement learning (RL) has garnered increasing recognition for its potential to optimise dynamic treatment regimes (DTRs) in personalised medicine.
We introduce textitDTR-Bench, a benchmarking platform for simulating diverse healthcare scenarios.
We evaluate various state-of-the-art RL algorithms across these settings, particularly highlighting their performance amidst real-world challenges.
arXiv Detail & Related papers (2024-05-28T21:40:00Z) - Self-STORM: Deep Unrolled Self-Supervised Learning for Super-Resolution Microscopy [55.2480439325792]
We introduce deep unrolled self-supervised learning, which alleviates the need for such data by training a sequence-specific, model-based autoencoder.
Our proposed method exceeds the performance of its supervised counterparts.
arXiv Detail & Related papers (2024-03-25T17:40:32Z) - Fast Cerebral Blood Flow Analysis via Extreme Learning Machine [4.373558495838564]
We introduce a rapid and precise analytical approach for analyzing cerebral blood flow (CBF) using Diffuse Correlation spectroscopy (DCS)
We assess existing algorithms using synthetic datasets for both semi-infinite and multi-layer models.
Results demonstrate that ELM consistently achieves higher fidelity across various noise levels and optical parameters, showcasing robust generalization ability and outperforming iterative fitting algorithms.
arXiv Detail & Related papers (2024-01-10T23:01:35Z) - Auto-FedRL: Federated Hyperparameter Optimization for
Multi-institutional Medical Image Segmentation [48.821062916381685]
Federated learning (FL) is a distributed machine learning technique that enables collaborative model training while avoiding explicit data sharing.
In this work, we propose an efficient reinforcement learning(RL)-based federated hyperparameter optimization algorithm, termed Auto-FedRL.
The effectiveness of the proposed method is validated on a heterogeneous data split of the CIFAR-10 dataset and two real-world medical image segmentation datasets.
arXiv Detail & Related papers (2022-03-12T04:11:42Z) - Learning Neural Causal Models with Active Interventions [83.44636110899742]
We introduce an active intervention-targeting mechanism which enables a quick identification of the underlying causal structure of the data-generating process.
Our method significantly reduces the required number of interactions compared with random intervention targeting.
We demonstrate superior performance on multiple benchmarks from simulated to real-world data.
arXiv Detail & Related papers (2021-09-06T13:10:37Z) - Dynamic Mode Decomposition in Adaptive Mesh Refinement and Coarsening
Simulations [58.720142291102135]
Dynamic Mode Decomposition (DMD) is a powerful data-driven method used to extract coherent schemes.
This paper proposes a strategy to enable DMD to extract from observations with different mesh topologies and dimensions.
arXiv Detail & Related papers (2021-04-28T22:14:25Z) - Deep Reinforcement Learning using Cyclical Learning Rates [62.19441737665902]
One of the most influential parameters in optimization procedures based on gradient descent (SGD) is the learning rate.
We investigate cyclical learning and propose a method for defining a general cyclical learning rate for various DRL problems.
Our experiments show that, utilizing cyclical learning achieves similar or even better results than highly tuned fixed learning rates.
arXiv Detail & Related papers (2020-07-31T10:06:02Z) - Longitudinal Deep Kernel Gaussian Process Regression [16.618767289437905]
We introduce Longitudinal deep kernel process regression (L-DKGPR)
L-DKGPR automates the discovery of complex multilevel correlation structure from longitudinal data.
We derive an efficient algorithm to train L-DKGPR using latent space inducing points and variational inference.
arXiv Detail & Related papers (2020-05-24T15:10:48Z) - A Systematic Approach to Featurization for Cancer Drug Sensitivity
Predictions with Deep Learning [49.86828302591469]
We train >35,000 neural network models, sweeping over common featurization techniques.
We found the RNA-seq to be highly redundant and informative even with subsets larger than 128 features.
arXiv Detail & Related papers (2020-04-30T20:42:17Z) - Distributed Training of Deep Neural Network Acoustic Models for
Automatic Speech Recognition [33.032361181388886]
We provide an overview of distributed training techniques for deep neural network acoustic models for ASR.
Experiments are carried out on a popular public benchmark to study the convergence, speedup and recognition performance of the investigated strategies.
arXiv Detail & Related papers (2020-02-24T19:31:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.