IntelligentPooling: Practical Thompson Sampling for mHealth
- URL: http://arxiv.org/abs/2008.01571v2
- Date: Sat, 12 Dec 2020 21:30:05 GMT
- Title: IntelligentPooling: Practical Thompson Sampling for mHealth
- Authors: Sabina Tomkins, Peng Liao, Predrag Klasnja and Susan Murphy
- Abstract summary: Reinforcement learning is ideal for learning how to optimally make sequential treatment decisions.
We generalize Thompson-Sampling bandit algorithms to develop IntelligentPooling.
We show that IntelligentPooling achieves an average of 26% lower regret than state-of-the-art.
- Score: 0.38998241153792446
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In mobile health (mHealth) smart devices deliver behavioral treatments
repeatedly over time to a user with the goal of helping the user adopt and
maintain healthy behaviors. Reinforcement learning appears ideal for learning
how to optimally make these sequential treatment decisions. However,
significant challenges must be overcome before reinforcement learning can be
effectively deployed in a mobile healthcare setting. In this work we are
concerned with the following challenges: 1) individuals who are in the same
context can exhibit differential response to treatments 2) only a limited
amount of data is available for learning on any one individual, and 3)
non-stationary responses to treatment. To address these challenges we
generalize Thompson-Sampling bandit algorithms to develop IntelligentPooling.
IntelligentPooling learns personalized treatment policies thus addressing
challenge one. To address the second challenge, IntelligentPooling updates each
user's degree of personalization while making use of available data on other
users to speed up learning. Lastly, IntelligentPooling allows responsivity to
vary as a function of a user's time since beginning treatment, thus addressing
challenge three. We show that IntelligentPooling achieves an average of 26%
lower regret than state-of-the-art. We demonstrate the promise of this approach
and its ability to learn from even a small group of users in a live clinical
trial.
Related papers
- Can machine learning solve the challenge of adaptive learning and the individualization of learning paths? A field experiment in an online learning platform [0.8437187555622164]
The individualization of learning contents based on digital technologies promises large individual and social benefits.
We conduct a randomized controlled trial on a large digital self-learning platform.
We develop an algorithm based on two convolutional neural networks that assigns tasks to $4,365$ learners according to their learning paths.
arXiv Detail & Related papers (2024-07-03T14:04:05Z) - Learning Task Decomposition to Assist Humans in Competitive Programming [90.4846613669734]
We introduce a novel objective for learning task decomposition, termed value (AssistV)
We collect a dataset of human repair experiences on different decomposed solutions.
Under 177 hours of human study, our method enables non-experts to solve 33.3% more problems, speeds them up by 3.3x, and empowers them to match unassisted experts.
arXiv Detail & Related papers (2024-06-07T03:27:51Z) - Uncertainty of Thoughts: Uncertainty-Aware Planning Enhances Information Seeking in Large Language Models [73.79091519226026]
Uncertainty of Thoughts (UoT) is an algorithm to augment large language models with the ability to actively seek information by asking effective questions.
In experiments on medical diagnosis, troubleshooting, and the 20 Questions game, UoT achieves an average performance improvement of 38.1% in the rate of successful task completion.
arXiv Detail & Related papers (2024-02-05T18:28:44Z) - Determining the Difficulties of Students With Dyslexia via Virtual
Reality and Artificial Intelligence: An Exploratory Analysis [0.0]
The VRAIlexia project has been created to tackle this issue by proposing two different tools.
The first one has been created and is being distributed among dyslexic students in Higher Education Institutions, for the conduction of specific psychological and psychometric tests.
The second tool applies specific artificial intelligence algorithms to the data gathered via the application and other surveys.
arXiv Detail & Related papers (2024-01-15T20:26:09Z) - BrainWash: A Poisoning Attack to Forget in Continual Learning [22.512552596310176]
"BrainWash" is a novel data poisoning method tailored to impose forgetting on a continual learner.
An important feature of our approach is that the attacker requires no access to previous tasks' data.
Our experiments highlight the efficacy of BrainWash, showcasing degradation in performance across various regularization-based continual learning methods.
arXiv Detail & Related papers (2023-11-20T18:26:01Z) - Optimising Human-AI Collaboration by Learning Convincing Explanations [62.81395661556852]
We propose a method for a collaborative system that remains safe by having a human making decisions.
Ardent enables efficient and effective decision-making by adapting to individual preferences for explanations.
arXiv Detail & Related papers (2023-11-13T16:00:16Z) - Resilient Constrained Learning [94.27081585149836]
This paper presents a constrained learning approach that adapts the requirements while simultaneously solving the learning task.
We call this approach resilient constrained learning after the term used to describe ecological systems that adapt to disruptions by modifying their operation.
arXiv Detail & Related papers (2023-06-04T18:14:18Z) - Continual Learning and Private Unlearning [49.848423659220444]
This paper formalizes the continual learning and private unlearning (CLPU) problem.
It introduces a straightforward but exactly private solution, CLPU-DER++, as the first step towards solving the CLPU problem.
arXiv Detail & Related papers (2022-03-24T02:40:33Z) - PEBBLE: Feedback-Efficient Interactive Reinforcement Learning via
Relabeling Experience and Unsupervised Pre-training [94.87393610927812]
We present an off-policy, interactive reinforcement learning algorithm that capitalizes on the strengths of both feedback and off-policy learning.
We demonstrate that our approach is capable of learning tasks of higher complexity than previously considered by human-in-the-loop methods.
arXiv Detail & Related papers (2021-06-09T14:10:50Z) - Guided Exploration with Proximal Policy Optimization using a Single
Demonstration [5.076419064097734]
We train an agent on a combination of demonstrations and own experience to solve problems with variable initial conditions.
The agent is able to increase its performance and to tackle harder problems by replaying its own past trajectories.
To the best of our knowledge, learning a task in a three-dimensional environment with comparable difficulty has never been considered before using only one human demonstration.
arXiv Detail & Related papers (2020-07-07T10:30:32Z) - Rapidly Personalizing Mobile Health Treatment Policies with Limited Data [9.07325490998379]
We present IntelligentPooling, which learns personalized policies via an adaptive, principled use of other users' data.
We show that IntelligentPooling achieves an average of 26% lower regret than state-of-the-art across all generative models.
arXiv Detail & Related papers (2020-02-23T18:59:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.