Character Ai Issues Apology After Teen Commits Suicide Over Ai Chatbot Obsession


The New York Times reports that the ninth grader started interacting with a chatbot named “Dany”, modeled on Daenerys Targaryen from the series “Game of Thrones”. He frequently shared personal information and role-played with the AI character, often indulging in romantic or sexual conversations.
The obsession grew so much that he preferred interacting with the AI character over real people, and his schoolwork suffered, as a result. After observing these behavioral changes, his parents took him to a therapist. He was diagnosed with anxiety and disruptive mood dysregulation disorder.
In his journal, he wrote, “I like staying in my room so much because I start to detach from this ‘reality,’ and I also feel more at peace, more connected with Dany and much more in love with her, and just happier.“
Later, the teenager expresses feelings of self-hatred and emptiness and tells the AI character ‘Dany’ that he loves her and will soon “come home” to her. He then ended his life. Now, the teenager’s mother, Megan L. Garcia, has filed a lawsuit against Character.ai, holding the company accountable for her son’s death. Character.ai has offered its deepest condolences to the family calling it the “tragic loss of one of our users.“
Character.ai has more than 20 million users, and most of them are young. The company says it takes the safety of users seriously. It has developed a pop-up that takes the user to a suicide prevention hotline whenever self-harm-related keywords are detected. However, this safety feature was not deployed when the teenager ended his life.
Character.ai allows minors who are at least 13 to use its services in the US. In fact, the service markets itself as a one-stop platform where you can “feel alive” and chat with an AI “Psychologist” and discuss life problems.
This particular case raises serious questions about AI companions and its impact on young users. We hope the lawsuit leads to stringent safety guardrails on AI platforms.