Learning-Based Strategy Design for Robot-Assisted Reminiscence Therapy
Based on a Developed Model for People with Dementia
- URL: http://arxiv.org/abs/2109.02194v1
- Date: Mon, 6 Sep 2021 00:45:31 GMT
- Title: Learning-Based Strategy Design for Robot-Assisted Reminiscence Therapy
Based on a Developed Model for People with Dementia
- Authors: Fengpei Yuan, Ran Zhang, Dania Bilal and Xiaopeng Zhao
- Abstract summary: The robot-assisted Reminiscence Therapy (RT) is studied as a psychosocial intervention to persons with dementia (PwDs)
We aim at a conversation strategy for the robot by learning to stimulate the PwD to talk.
- Score: 2.453923815224908
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, the robot-assisted Reminiscence Therapy (RT) is studied as a
psychosocial intervention to persons with dementia (PwDs). We aim at a
conversation strategy for the robot by reinforcement learning to stimulate the
PwD to talk. Specifically, to characterize the stochastic reactions of a PwD to
the robot's actions, a simulation model of a PwD is developed which features
the transition probabilities among different PwD states consisting of the
response relevance, emotion levels and confusion conditions. A Q-learning (QL)
algorithm is then designed to achieve the best conversation strategy for the
robot. The objective is to stimulate the PwD to talk as much as possible while
keeping the PwD's states as positive as possible. In certain conditions, the
achieved strategy gives the PwD choices to continue or change the topic, or
stop the conversation, so that the PwD has a sense of control to mitigate the
conversation stress. To achieve this, the standard QL algorithm is revised to
deliberately integrate the impact of PwD's choices into the Q-value updates.
Finally, the simulation results demonstrate the learning convergence and
validate the efficacy of the achieved strategy. Tests show that the strategy is
capable to duly adjust the difficulty level of prompt according to the PwD's
states, take actions (e.g., repeat or explain the prompt, or comfort) to help
the PwD out of bad states, and allow the PwD to control the conversation
tendency when bad states continue.
Related papers
- Interactive Dialogue Agents via Reinforcement Learning on Hindsight Regenerations [58.65755268815283]
Many real dialogues are interactive, meaning an agent's utterances will influence their conversational partner, elicit information, or change their opinion.
We use this fact to rewrite and augment existing suboptimal data, and train via offline reinforcement learning (RL) an agent that outperforms both prompting and learning from unaltered human demonstrations.
Our results in a user study with real humans show that our approach greatly outperforms existing state-of-the-art dialogue agents.
arXiv Detail & Related papers (2024-11-07T21:37:51Z) - Prompt Engineering a Schizophrenia Chatbot: Utilizing a Multi-Agent Approach for Enhanced Compliance with Prompt Instructions [0.0699049312989311]
Patients with schizophrenia often present with cognitive impairments that may hinder their ability to learn about their condition.
While Large Language Models (LLMs) have the potential to make topical mental health information more accessible and engaging, their black-box nature raises concerns about ethics and safety.
arXiv Detail & Related papers (2024-10-10T09:49:24Z) - Dialogue Action Tokens: Steering Language Models in Goal-Directed Dialogue with a Multi-Turn Planner [51.77263363285369]
We present an approach called Dialogue Action Tokens that adapts language model agents to plan goal-directed dialogues.
The core idea is to treat each utterance as an action, thereby converting dialogues into games where existing approaches such as reinforcement learning can be applied.
arXiv Detail & Related papers (2024-06-17T18:01:32Z) - Social Motion Prediction with Cognitive Hierarchies [19.71780279070757]
We introduce a new benchmark, a novel formulation, and a cognition-inspired framework.
We present Wusi, a 3D multi-person motion dataset under the context of team sports.
We develop a cognitive hierarchy framework to predict strategic human social interactions.
arXiv Detail & Related papers (2023-11-08T14:51:17Z) - Harnessing Large Language Models' Empathetic Response Generation
Capabilities for Online Mental Health Counselling Support [1.9336815376402723]
Large Language Models (LLMs) have demonstrated remarkable performance across various information-seeking and reasoning tasks.
This study sought to examine LLMs' capability to generate empathetic responses in conversations that emulate those in a mental health counselling setting.
We selected five LLMs: version 3.5 and version 4 of the Generative Pre-training (GPT), Vicuna FastChat-T5, Pathways Language Model (PaLM) version 2, and Falcon-7B-Instruct.
arXiv Detail & Related papers (2023-10-12T03:33:06Z) - Reflective Linguistic Programming (RLP): A Stepping Stone in
Socially-Aware AGI (SocialAGI) [0.0]
This paper presents Reflective Linguistic Programming (RLP), a unique approach to conversational AI that emphasizes self-awareness and strategic planning.
RLP encourages models to introspect on their own predefined personality traits, emotional responses to incoming messages, and planned strategies, enabling contextually rich, coherent, and engaging interactions.
arXiv Detail & Related papers (2023-05-22T02:43:15Z) - Controllable Mixed-Initiative Dialogue Generation through Prompting [50.03458333265885]
Mixed-initiative dialogue tasks involve repeated exchanges of information and conversational control.
Agents gain control by generating responses that follow particular dialogue intents or strategies, prescribed by a policy planner.
Standard approach has been fine-tuning pre-trained language models to perform generation conditioned on these intents.
We instead prompt large language models as a drop-in replacement to fine-tuning on conditional generation.
arXiv Detail & Related papers (2023-05-06T23:11:25Z) - Achieving mouse-level strategic evasion performance using real-time
computational planning [59.60094442546867]
Planning is an extraordinary ability in which the brain imagines and then enacts evaluated possible futures.
We develop a more efficient biologically-inspired planning algorithm, TLPPO, based on work on how the ecology of an animal governs the value of spatial planning.
We compare the performance of a real-time agent using TLPPO against the performance of live mice, all tasked with evading a robot predator.
arXiv Detail & Related papers (2022-11-04T18:34:36Z) - A Simulated Experiment to Explore Robotic Dialogue Strategies for People
with Dementia [2.5412519393131974]
We propose a partially observable markov decision process (POMDP) model for the PwD-robot interaction in the context of repetitive questioning.
We used Q-learning to learn an adaptive conversation strategy towards PwDs with different cognitive capabilities and different engagement levels.
This may be a useful step towards the application of conversational social robots to cope with repetitive questioning in PwDs.
arXiv Detail & Related papers (2021-04-18T19:35:19Z) - Evolutionary Stochastic Policy Distillation [139.54121001226451]
We propose a new method called Evolutionary Policy Distillation (ESPD) to solve GCRS tasks.
ESPD enables a target policy to learn from a series of its variants through the technique of policy distillation (PD)
The experiments based on the MuJoCo control suite show the high learning efficiency of the proposed method.
arXiv Detail & Related papers (2020-04-27T16:19:25Z) - You Impress Me: Dialogue Generation via Mutual Persona Perception [62.89449096369027]
The research in cognitive science suggests that understanding is an essential signal for a high-quality chit-chat conversation.
Motivated by this, we propose P2 Bot, a transmitter-receiver based framework with the aim of explicitly modeling understanding.
arXiv Detail & Related papers (2020-04-11T12:51:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.