Adaptive Interventions with User-Defined Goals for Health Behavior Change
- URL: http://arxiv.org/abs/2311.09483v4
- Date: Thu, 23 May 2024 15:56:04 GMT
- Title: Adaptive Interventions with User-Defined Goals for Health Behavior Change
- Authors: Aishwarya Mandyam, Matthew Jörke, William Denton, Barbara E. Engelhardt, Emma Brunskill,
- Abstract summary: Mobile health applications present a promising avenue for low-cost, scalable health behavior change promotion.
tailoring advice to a person's unique goals, preferences, and life circumstances is a critical component of health coaching.
We introduce a new Thompson sampling algorithm that can accommodate personalized reward functions.
- Score: 17.688448640253494
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Promoting healthy lifestyle behaviors remains a major public health concern, particularly due to their crucial role in preventing chronic conditions such as cancer, heart disease, and type 2 diabetes. Mobile health applications present a promising avenue for low-cost, scalable health behavior change promotion. Researchers are increasingly exploring adaptive algorithms that personalize interventions to each person's unique context. However, in empirical studies, mobile health applications often suffer from small effect sizes and low adherence rates, particularly in comparison to human coaching. Tailoring advice to a person's unique goals, preferences, and life circumstances is a critical component of health coaching that has been underutilized in adaptive algorithms for mobile health interventions. To address this, we introduce a new Thompson sampling algorithm that can accommodate personalized reward functions (i.e., goals, preferences, and constraints), while also leveraging data sharing across individuals to more quickly be able to provide effective recommendations. We prove that our modification incurs only a constant penalty on cumulative regret while preserving the sample complexity benefits of data sharing. We present empirical results on synthetic and semi-synthetic physical activity simulators, where in the latter we conducted an online survey to solicit preference data relating to physical activity, which we use to construct realistic reward models that leverages historical data from another study. Our algorithm achieves substantial performance improvements compared to baselines that do not share data or do not optimize for individualized rewards.
Related papers
- Reduced-Rank Multi-objective Policy Learning and Optimization [57.978477569678844]
In practice, causal researchers do not have a single outcome in mind a priori.
In government-assisted social benefit programs, policymakers collect many outcomes to understand the multidimensional nature of poverty.
We present a data-driven dimensionality-reduction methodology for multiple outcomes in the context of optimal policy learning.
arXiv Detail & Related papers (2024-04-29T08:16:30Z) - Generating Synthetic Health Sensor Data for Privacy-Preserving Wearable Stress Detection [1.3604778572442302]
We introduce the privacy-aware synthetization of multi-sensor smartwatch health readings related to moments of stress.
Our method not only protects patient information but also enhances data availability for research.
arXiv Detail & Related papers (2024-01-24T09:44:57Z) - Policy Optimization for Personalized Interventions in Behavioral Health [8.10897203067601]
Behavioral health interventions, delivered through digital platforms, have the potential to significantly improve health outcomes.
We study the problem of optimizing personalized interventions for patients to maximize a long-term outcome.
We present a new approach for this problem that we dub DecompPI, which decomposes the state space for a system of patients to the individual level.
arXiv Detail & Related papers (2023-03-21T21:42:03Z) - Zero-shot causal learning [64.9368337542558]
CaML is a causal meta-learning framework which formulates the personalized prediction of each intervention's effect as a task.
We show that CaML is able to predict the personalized effects of novel interventions that do not exist at the time of training.
arXiv Detail & Related papers (2023-01-28T20:14:11Z) - Data-pooling Reinforcement Learning for Personalized Healthcare
Intervention [20.436521180168455]
We develop a novel data-pooling reinforcement learning (RL) algorithm based on a general perturbed value iteration framework.
Our algorithm adaptively pools historical data, with three main innovations: (i) the weight of pooling ties directly to the performance of decision (measured by regret) as opposed to estimation accuracy in conventional methods.
We substantiate the theoretical development with empirically better performance of our algorithm via a case study in the context of post-discharge intervention to prevent unplanned readmissions.
arXiv Detail & Related papers (2022-11-16T15:52:49Z) - Adaptive Identification of Populations with Treatment Benefit in
Clinical Trials: Machine Learning Challenges and Solutions [78.31410227443102]
We study the problem of adaptively identifying patient subpopulations that benefit from a given treatment during a confirmatory clinical trial.
We propose AdaGGI and AdaGCPI, two meta-algorithms for subpopulation construction.
arXiv Detail & Related papers (2022-08-11T14:27:49Z) - Practical Challenges in Differentially-Private Federated Survival
Analysis of Medical Data [57.19441629270029]
In this paper, we take advantage of the inherent properties of neural networks to federate the process of training of survival analysis models.
In the realistic setting of small medical datasets and only a few data centers, this noise makes it harder for the models to converge.
We propose DPFed-post which adds a post-processing stage to the private federated learning scheme.
arXiv Detail & Related papers (2022-02-08T10:03:24Z) - DTR Bandit: Learning to Make Response-Adaptive Decisions With Low Regret [59.81290762273153]
Dynamic treatment regimes (DTRs) are personalized, adaptive, multi-stage treatment plans that adapt treatment decisions to an individual's initial features and to intermediate outcomes and features at each subsequent stage.
We propose a novel algorithm that, by carefully balancing exploration and exploitation, is guaranteed to achieve rate-optimal regret when the transition and reward models are linear.
arXiv Detail & Related papers (2020-05-06T13:03:42Z) - Power Constrained Bandits [46.44025793243983]
We develop general meta-algorithms to modify existing algorithms.
Our meta-algorithms are robust to various model mis-specifications possibly appearing in statistical studies.
arXiv Detail & Related papers (2020-04-13T23:08:52Z) - Rapidly Personalizing Mobile Health Treatment Policies with Limited Data [9.07325490998379]
We present IntelligentPooling, which learns personalized policies via an adaptive, principled use of other users' data.
We show that IntelligentPooling achieves an average of 26% lower regret than state-of-the-art across all generative models.
arXiv Detail & Related papers (2020-02-23T18:59:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.