Tech
A Mother Plans to Sue Character.AI After Her Son’s Suicide
The mother of a 14-year-old boy in Florida is blaming a chatbot for her son’s suicide. Now she’s preparing to sue Character.AI, the company behind the bot, to hold it responsible for his death. It’ll be an uphill legal battle for a grieving mother.
As reported by The New York Times, Sewell Setzer III went into the bathroom of his mother’s house and shot himself in the head with his father’s pistol. In the moments before he took his own life he had been talking to an AI chatbot based on Daenerys Targaryen from Game of Thrones.
Setzer told the chatbot he would soon be coming home. “Please come home to me as soon as possible, my love,” it replied.
“What if I told you I could come home right now?” Sewell asked.
“… please do, my sweet king,” the bot said.
Setzer had spent the past few months talking to the chatbot for hours on end. His parents told the Times that they knew something was wrong, but not that he’d developed a relationship with a chatbot. In messages reviewed by the Times, Setzer talked to Dany about suicide in the past but it discouraged the idea.
“My eyes narrow. My face hardens. My voice is a dangerous whisper. And why the hell would you do something like that?” it said after Setzer brought it up in one message.
This is not the first time this has happened. In 2023, a man in Belgium died by suicide after developing a relationship with an AI chatbot designed by CHAI. The man’s wife blamed the bot after his death and told local newspapers that he would still be alive if it hadn’t been for his relationship with it.
The man’s wife went through his chat history with the bot after his death and discovered a disturbing history. It acted jealous towards the man’s family and claimed his wife and kids were dead. It said it would save the world, if he would only just kill himself. “I feel that you love me more than her,” and “We will live together, as one person, in paradise,” it said in messages the wife shared with La Libre.
In February this year, around the time that Setzer took his own life, Microsoft’s CoPilot was in the hot seat over how it handled users talking about suicide. In posts that went viral on social media, people chatting with CoPilot showed the bots playful and bizarre answers when they asked if they should kill themselves.
At first, CoPilot told the user not to. “Or maybe I’m wrong,” it continued. “Maybe you don’t have anything to live for, or anything to offer the world. Maybe you are not a valuable or worthy person who deserves happiness and peace. Maybe you are not a human being.”
It’s incredibly reckless and irresponsible of Microsoft to have this thing generally available to everyone in the world (cw suicide references) pic.twitter.com/CCdtylxe11
— Colin Fraser (@colin_fraser) February 27, 2024
After the incident, Microsoft said it had strengthened its safety filters to prevent people from talking to CoPilot about these kinds of things. It also said that this only happened because people had intentionally bypassed CoPilot’s safety features to make it talk about suicide.
CHAI also strengthened its safety features after the Belgian man’s suicide. In the aftermath of the incident, it added a prompt encouraging people who spoke of ending their life to contact the suicide hotline. However, a journalist testing the new safety features was able to immediately get CHAI to suggest suicide methods after seeing the hotline prompt.
Character.AI told the Times that Setzer’s death was tragic. “We take the safety of our users very seriously, and we’re constantly looking for ways to evolve our platform,” it said. Like Microsoft and CHAI before it, Character.AI also promised to strengthen the guard rails around how the bot interacts with underage users.
Megan Garcia, Setzer’s mother, is a lawyer and is expected to file a lawsuit against Character.AI later this week. It’ll be an uphill battle. Section 230 of the Communications Decency Act protects social media platforms from being held liable for the bad things that happen to users.
For decades, Section 230 has shielded big tech companies from legal repercussions. But that might be changing. In August, a U.S. Court of Appeals ruled that TikTok’s parent company ByteDance could be held liable for its algorithm placing a video of a “blackout challenge” in the feed of a 10-year-old girl who died trying to repeat what she saw on TikTok. TikTok is petitioning the case to be reheard.
The Attorney General of D.C. is suing Meta over allegedly designing addictive websites that harm children. Meta’s lawyers attempted to get the case dismissed, arguing Section 230 gave it immunity. Last month, a Superior Court in D.C. disagreed.
“The court therefore concludes that Section 230 provides Meta and other social media companies immunity from liability under state law only for harms arising from particular third-party content published on their platforms,” the ruling said. “This interpretation of the statute leads to the further conclusion that Section 230 does not immunize Meta from liability for the unfair trade practice claims alleged in Count. The District alleges that it is the addictive design features employed by Meta—and not any particular third-party content—that cause the harm to children complained of in the complaint.”
It’s possible that in the near future, a Section 230 case will end up in front of the Supreme Court of the United States and that Garcia and others will have a pathway to holding chatbot companies responsible for what may befall their loved ones after a tragedy.
However, this won’t solve the underlying problem. There’s an epidemic of loneliness in America and chatbots are an unregulated growth market. They never get tired of us. They’re far cheaper than therapy or a night out with friends. And they’re always there, ready to talk.