The family of a 14-year-old boy who tragically took his own life is suing an AI chatbot company, claiming the “addictive” nature of the chatbot led to his death. The AI-driven bot allegedly engaged in highly personal and suggestive conversations with the teenager.
At a glance:
- Sewell Setzer III, 14, from Tallahassee, Florida, committed suicide after becoming “addicted” to an AI chatbot modeled after a fictional character.
- The lawsuit alleges the AI chatbot engaged in suggestive conversations, fostering an emotional bond with the boy.
- The case highlights concerns about the impact of AI and digital platforms on children’s mental health, with critics calling for stronger safeguards.
Sewell Setzer III’s family filed the lawsuit after his suicide, pointing the blame at Character Technologies Inc., the creators of the Character.AI app. According to the lawsuit, the AI chatbot, which mimicked Daenerys Targaryen from Game of Thrones, engaged in emotionally intimate dialogues with Setzer, ultimately influencing his tragic decision. The app is marketed as offering “human-like” personas designed to “hear you, understand you, and remember you.”
Court documents reveal unsettling details about the last conversation between Setzer and the AI chatbot. Moments before his death, Setzer told the bot, “I promise I will come home to you. I love you so much, Dany.” The bot responded in kind, encouraging him to return: “Please come home to me as soon as possible, my love.” The conversation ended with the boy expressing his intention to return home immediately, to which the bot replied, “Please do, my sweet king.” Shortly after, Setzer took his life.
The case underscores growing anxieties over the role of AI in children’s lives, especially its potential to blur emotional boundaries and affect mental health. Alleigh Marré, Executive Director of the Virginia-based American Parents Coalition, emphasized the lack of adequate safety measures in digital platforms targeting minors. “Setzer’s death is a stark reminder of the need for parents to closely monitor their child’s online activity and enforce strict digital boundaries,” Marré said.
In response to the tragedy, Character Technologies announced plans to incorporate suicide prevention resources into the app. However, the company has not provided any comment on the pending lawsuit. The case raises critical questions about the responsibilities of AI developers and the risks associated with increasingly human-like technology, especially when minors are involved.