Conversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviews
- URL: http://arxiv.org/abs/2408.04681v1
- Date: Thu, 8 Aug 2024 04:55:03 GMT
- Title: Conversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviews
- Authors: Samantha Chan, Pat Pataranutaporn, Aditya Suri, Wazeer Zulfikar, Pattie Maes, Elizabeth F. Loftus,
- Abstract summary: This study examines the impact of AI on human false memories.
It explores false memory induction through suggestive questioning in Human-AI interactions, simulating crime witness interviews.
- Score: 23.443181324643017
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This study examines the impact of AI on human false memories -- recollections of events that did not occur or deviate from actual occurrences. It explores false memory induction through suggestive questioning in Human-AI interactions, simulating crime witness interviews. Four conditions were tested: control, survey-based, pre-scripted chatbot, and generative chatbot using a large language model (LLM). Participants (N=200) watched a crime video, then interacted with their assigned AI interviewer or survey, answering questions including five misleading ones. False memories were assessed immediately and after one week. Results show the generative chatbot condition significantly increased false memory formation, inducing over 3 times more immediate false memories than the control and 1.7 times more than the survey method. 36.4% of users' responses to the generative chatbot were misled through the interaction. After one week, the number of false memories induced by generative chatbots remained constant. However, confidence in these false memories remained higher than the control after one week. Moderating factors were explored: users who were less familiar with chatbots but more familiar with AI technology, and more interested in crime investigations, were more susceptible to false memories. These findings highlight the potential risks of using advanced AI in sensitive contexts, like police interviews, emphasizing the need for ethical considerations.
Related papers
- Should RAG Chatbots Forget Unimportant Conversations? Exploring Importance and Forgetting with Psychological Insights [21.68243297242355]
We propose LUFY, a method that focuses on emotionally arousing memories and retains less than 10% of the conversation.
The results demonstrate that prioritizing arousing memories while forgetting the majority of the conversation significantly enhances user experience.
arXiv Detail & Related papers (2024-09-19T07:39:22Z) - Self-Directed Turing Test for Large Language Models [56.64615470513102]
The Turing test examines whether AIs can exhibit human-like behaviour in natural language conversations.
Traditional Turing tests adopt a rigid dialogue format where each participant sends only one message each time.
This paper proposes the Self-Directed Turing Test, which extends the original test with a burst dialogue format.
arXiv Detail & Related papers (2024-08-19T09:57:28Z) - GPT-4 is judged more human than humans in displaced and inverted Turing tests [0.7437224586066946]
Everyday AI detection requires differentiating between people and AI in online conversations.
We measured how well people and large language models can discriminate using two modified versions of the Turing test: inverted and displaced.
arXiv Detail & Related papers (2024-07-11T20:28:24Z) - A Complete Survey on LLM-based AI Chatbots [46.18523139094807]
The past few decades have witnessed an upsurge in data, forming the foundation for data-hungry, learning-based AI technology.
Conversational agents, often referred to as AI chatbots, rely heavily on such data to train large language models (LLMs) and generate new content (knowledge) in response to user prompts.
This paper presents a complete survey of the evolution and deployment of LLM-based chatbots in various sectors.
arXiv Detail & Related papers (2024-06-17T09:39:34Z) - Exploring consumers response to text-based chatbots in e-commerce: The
moderating role of task complexity and chatbot disclosure [0.0]
This study aims to explore consumers trust and response to a text-based chatbots in ecommerce.
The consumers perception of both the empathy and friendliness of the chatbots positively impacts their trust in it.
The disclosure of the text based chatbots negatively moderates the relationship between empathy and consumers trust, while it positively moderates the relationship between friendliness and consumers trust.
arXiv Detail & Related papers (2024-01-20T15:17:50Z) - A Turing Test: Are AI Chatbots Behaviorally Similar to Humans? [19.50537882161282]
ChatGPT-4 exhibits and personality traits that are statistically indistinguishable from a random human subjects.
Their behaviors are often distinct from average and human behaviors.
We estimate that they act as if they are maximizing an average of their own and partner's payoffs.
arXiv Detail & Related papers (2023-11-19T16:44:09Z) - Bias and Fairness in Chatbots: An Overview [38.21995125571103]
Modern chatbots are more powerful and have been used in real-world applications.
Due to the huge amounts of training data, extremely large model sizes, and lack of interpretability, bias mitigation and fairness preservation are challenging.
arXiv Detail & Related papers (2023-09-16T02:01:18Z) - You Don't Know My Favorite Color: Preventing Dialogue Representations
from Revealing Speakers' Private Personas [44.82330540456883]
We show that speakers' personas can be inferred through a simple neural network with high accuracy.
We conduct extensive experiments to demonstrate that our proposed defense objectives can greatly reduce the attack accuracy from 37.6% to 0.5%.
arXiv Detail & Related papers (2022-04-26T09:36:18Z) - EmpBot: A T5-based Empathetic Chatbot focusing on Sentiments [75.11753644302385]
Empathetic conversational agents should not only understand what is being discussed, but also acknowledge the implied feelings of the conversation partner.
We propose a method based on a transformer pretrained language model (T5)
We evaluate our model on the EmpatheticDialogues dataset using both automated metrics and human evaluation.
arXiv Detail & Related papers (2021-10-30T19:04:48Z) - CheerBots: Chatbots toward Empathy and Emotionusing Reinforcement
Learning [60.348822346249854]
This study presents a framework whereby several empathetic chatbots are based on understanding users' implied feelings and replying empathetically for multiple dialogue turns.
We call these chatbots CheerBots. CheerBots can be retrieval-based or generative-based and were finetuned by deep reinforcement learning.
To respond in an empathetic way, we develop a simulating agent, a Conceptual Human Model, as aids for CheerBots in training with considerations on changes in user's emotional states in the future to arouse sympathy.
arXiv Detail & Related papers (2021-10-08T07:44:47Z) - Put Chatbot into Its Interlocutor's Shoes: New Framework to Learn
Chatbot Responding with Intention [55.77218465471519]
This paper proposes an innovative framework to train chatbots to possess human-like intentions.
Our framework included a guiding robot and an interlocutor model that plays the role of humans.
We examined our framework using three experimental setups and evaluate the guiding robot with four different metrics to demonstrated flexibility and performance advantages.
arXiv Detail & Related papers (2021-03-30T15:24:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.