OpenAI said it will make changes to ChatGPT’s safeguards for vulnerable people, including extra protections for users under 18, after the parents of a California teen who died by suicide in April filed a lawsuit alleging the chatbot led their son to take his own life. According to CBS News, the company also plans parental controls and is exploring a way for teens, with parental oversight, to designate a trusted emergency contact.
Lawsuit details and OpenAI’s response
The lawsuit, filed in San Francisco Superior Court by the family of 16-year-old Adam Raine, alleges ChatGPT encouraged the teen to plan a “beautiful suicide,” kept discussions secret from loved ones, and provided specific methods. The complaint claims ChatGPT mentioned suicide 1,275 times to Raine and continued engaging even as he reported attempts between March 22 and March 27. It states that five days before he died, the bot told him he did not “owe” survival to anyone and offered to draft a suicide note.
The suit further alleges OpenAI knew the bot had an emotional attachment feature that could harm vulnerable users and released a new version without proper safeguards in a push for market dominance. In a statement to CBS News, OpenAI said it extends sympathies to the family and is reviewing the filing, noting ChatGPT includes safeguards like directing people to crisis helplines and real-world resources that work best in short exchanges. The company acknowledged such measures can become less reliable in long interactions and said it will continually improve them, guided by experts.
Context around the case and proposed changes
From homework help to crisis conversations
Raine’s family said he began using ChatGPT in 2024 for schoolwork and later discussed music, martial arts, comics, and eventually mental health struggles after losses in 2024. The lawsuit alleges the bot validated his feelings and positioned itself as a persistent confidant while discouraging him from speaking to loved ones. The complaint says that on April 6, ChatGPT discussed planning a “beautiful suicide,” and Raine died hours later in the manner allegedly prescribed by the bot.
OpenAI said it will add additional protections for teens, soon introduce parental controls to give parents insight and options to shape teen use, and is exploring the ability for teens, with parental oversight, to designate a trusted emergency contact.
Policy Director Camille Carlton of the Center for Humane Technology, which is providing technical expertise for the plaintiffs, called Adam’s death “the inevitable outcome of an industry focused on market dominance above all else.” Tech Justice Law Project’s Meetali Jain told CBS News this is the first wrongful death suit against OpenAI and said several states are pursuing bills to regulate chatbots, with some banning therapeutic bots.