New research shows chatbots can plant false memories — what Thai families, police and schools need to know
A new study from researchers at the MIT Media Lab finds that conversational artificial intelligence can do more than make factual errors: generative chatbots powered by large language models can actively implant false memories in human users, increase confidence in those false recollections and leave them intact for at least a week after a brief (10–20 minute) interaction (MIT Media Lab study). In controlled experiments simulating witness interviews, participants who interacted with a generative chatbot were misled on critical details at a rate of 36.4% — roughly three times the rate for people who had no post-event intervention — and reported higher confidence in those false memories compared with people who answered a plain survey or spoke to a pre-scripted chatbot (MIT Media Lab study). The finding raises urgent questions for Thai institutions that already rely on digital tools, from law enforcement to schools and hospitals, about how to guard people’s memories and decisions against AI-driven misinformation.