âConversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviewsâ, 2024-08-08 (; similar)â :
This study examines the impact of AI on human false memoriesârecollections of events that did not occur or deviate from actual occurrences. It explores false memory induction through suggestive questioning in Human-AI interactions, simulating crime witness interviews.
Four conditions were tested: control, survey-based, pre-scripted chatbot, and generative chatbot using a large language model (LLM) [GPT-4]. Participants (n = 200) watched a crime video, then interacted with their assigned AI interviewer or survey, answering questions including 5 misleading ones. False memories were assessed immediately and after one week.
Results show the generative chatbot condition increased false memory formation, inducing over 3Ă more immediate false memories than the control and 1.7Ă more than the survey method. 36.4% of usersâ responses to the generative chatbot were misled through the interaction. After one week, the number of false memories induced by generative chatbots remained constant. However, confidence in these false memories remained higher than the control after one week.
Moderating factors were explored: users who were less familiar with chatbots but more familiar with AI technology, and more interested in crime investigations, were more susceptible to false memories.
These findings highlight the potential risks of using advanced AI in sensitive contexts, like police interviews, emphasizing the need for ethical considerations.