ExOSITO: Explainable Off-Policy Learning with Side Information for Intensive Care Unit Blood Test Orders
- URL: http://arxiv.org/abs/2504.17277v1
- Date: Thu, 24 Apr 2025 06:07:14 GMT
- Title: ExOSITO: Explainable Off-Policy Learning with Side Information for Intensive Care Unit Blood Test Orders
- Authors: Zongliang Ji, Andre Carlos Kajdacsy-Balla Amaral, Anna Goldenberg, Rahul G. Krishnan,
- Abstract summary: This paper develops a novel method that combines off-policy learning with privileged information to identify the optimal set of ICU lab tests to order.<n>We pose this problem as a causal bandit trained using offline data and a reward function derived from clinically-approved rules.<n>The learned policy function provides interpretable clinical information and reduces costs without omitting any vital lab orders.
- Score: 13.090586660373775
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Ordering a minimal subset of lab tests for patients in the intensive care unit (ICU) can be challenging. Care teams must balance between ensuring the availability of the right information and reducing the clinical burden and costs associated with each lab test order. Most in-patient settings experience frequent over-ordering of lab tests, but are now aiming to reduce this burden on both hospital resources and the environment. This paper develops a novel method that combines off-policy learning with privileged information to identify the optimal set of ICU lab tests to order. Our approach, EXplainable Off-policy learning with Side Information for ICU blood Test Orders (ExOSITO) creates an interpretable assistive tool for clinicians to order lab tests by considering both the observed and predicted future status of each patient. We pose this problem as a causal bandit trained using offline data and a reward function derived from clinically-approved rules; we introduce a novel learning framework that integrates clinical knowledge with observational data to bridge the gap between the optimal and logging policies. The learned policy function provides interpretable clinical information and reduces costs without omitting any vital lab orders, outperforming both a physician's policy and prior approaches to this practical problem.
Related papers
- Deep Causal Behavioral Policy Learning: Applications to Healthcare [0.0]
We present a deep learning-based approach to studying dynamic clinical behavioral regimes in diverse non-randomized healthcare settings.
Our proposed methodology uses deep learning algorithms to learn the distribution of high-dimensional clinical action paths.
We propose a novel interpretation of a behavioral policy learned using the LCBM: that it is an efficient encoding of complex, often implicit, knowledge used to treat a patient.
arXiv Detail & Related papers (2025-03-05T18:24:58Z) - Leveraging Self-Supervised Learning Methods for Remote Screening of Subjects with Paroxysmal Atrial Fibrillation [6.158210490716697]
This study explores the application of Self-Supervised Learning (SSL) as a way to obtain preliminary results from clinical studies with limited cohorts.
We focus on an underexplored clinical task: screening subjects for Paroxysmal Atrial Fibrillation (P-AF) using remote monitoring, single-lead ECG signals captured during normal sinus rhythm.
We evaluate state-of-the-art SSL methods alongside supervised learning approaches, where SSL outperforms supervised learning in this task of interest.
arXiv Detail & Related papers (2025-03-04T13:42:38Z) - Development and Validation of Heparin Dosing Policies Using an Offline Reinforcement Learning Algorithm [0.7519918949973486]
This study proposes a reinforcement learning-based personalized optimal heparin dosing policy.
A batch-constrained policy was implemented to minimize out-of-distribution errors in an offline RL environment.
This research enhances heparin administration practices and establishes a precedent for the development of sophisticated decision-support tools in medicine.
arXiv Detail & Related papers (2024-09-24T05:20:38Z) - TrialBench: Multi-Modal Artificial Intelligence-Ready Clinical Trial Datasets [57.067409211231244]
This paper presents meticulously curated AIready datasets covering multi-modal data (e.g., drug molecule, disease code, text, categorical/numerical features) and 8 crucial prediction challenges in clinical trial design.
We provide basic validation methods for each task to ensure the datasets' usability and reliability.
We anticipate that the availability of such open-access datasets will catalyze the development of advanced AI approaches for clinical trial design.
arXiv Detail & Related papers (2024-06-30T09:13:10Z) - Measurement Scheduling for ICU Patients with Offline Reinforcement
Learning [16.07235754244993]
Studies show that 20-40% of lab tests ordered in the ICU are redundant and could be eliminated without compromising patient safety.
Prior work has leveraged offline reinforcement learning (Offline-RL) to find optimal policies for ordering lab tests based on patient information.
New ICU patient datasets have since been released, and various advancements have been made in Offline-RL methods.
arXiv Detail & Related papers (2024-02-12T00:22:47Z) - Self-Verification Improves Few-Shot Clinical Information Extraction [73.6905567014859]
Large language models (LLMs) have shown the potential to accelerate clinical curation via few-shot in-context learning.
They still struggle with issues regarding accuracy and interpretability, especially in mission-critical domains such as health.
Here, we explore a general mitigation framework using self-verification, which leverages the LLM to provide provenance for its own extraction and check its own outputs.
arXiv Detail & Related papers (2023-05-30T22:05:11Z) - SPeC: A Soft Prompt-Based Calibration on Performance Variability of
Large Language Model in Clinical Notes Summarization [50.01382938451978]
We introduce a model-agnostic pipeline that employs soft prompts to diminish variance while preserving the advantages of prompt-based summarization.
Experimental findings indicate that our method not only bolsters performance but also effectively curbs variance for various language models.
arXiv Detail & Related papers (2023-03-23T04:47:46Z) - Towards Real-World Applications of Personalized Anesthesia Using Policy
Constraint Q Learning for Propofol Infusion Control [11.07500653906137]
Policy Constraint Q-Learning (PCQL) is a data-driven reinforcement learning algorithm for solving the problem of learning anesthesia strategies on real clinical datasets.
PCQL is validated by extensive experiments on a real clinical anesthesia dataset.
arXiv Detail & Related papers (2023-03-17T10:05:20Z) - Retrieval-Augmented and Knowledge-Grounded Language Models for Faithful Clinical Medicine [68.7814360102644]
We propose the Re$3$Writer method with retrieval-augmented generation and knowledge-grounded reasoning.
We demonstrate the effectiveness of our method in generating patient discharge instructions.
arXiv Detail & Related papers (2022-10-23T16:34:39Z) - Federated Cycling (FedCy): Semi-supervised Federated Learning of
Surgical Phases [57.90226879210227]
FedCy is a semi-supervised learning (FSSL) method that combines FL and self-supervised learning to exploit a decentralized dataset of both labeled and unlabeled videos.
We demonstrate significant performance gains over state-of-the-art FSSL methods on the task of automatic recognition of surgical phases.
arXiv Detail & Related papers (2022-03-14T17:44:53Z) - HINT: Hierarchical Interaction Network for Trial Outcome Prediction
Leveraging Web Data [56.53715632642495]
Clinical trials face uncertain outcomes due to issues with efficacy, safety, or problems with patient recruitment.
In this paper, we propose Hierarchical INteraction Network (HINT) for more general, clinical trial outcome predictions.
arXiv Detail & Related papers (2021-02-08T15:09:07Z) - Contextual Constrained Learning for Dose-Finding Clinical Trials [102.8283665750281]
C3T-Budget is a contextual constrained clinical trial algorithm for dose-finding under both budget and safety constraints.
It recruits patients with consideration of the remaining budget, the remaining time, and the characteristics of each group.
arXiv Detail & Related papers (2020-01-08T11:46:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.