OpenAI CEO Sam Altman speaks during the Federal Reserve’s Integrated Review of the Capital Framework for Large Banks Conference in Washington, D.C., U.S., July 22, 2025.

Ken Cedeno | Reuters

OpenAI is detailing its plans to address ChatGPT’s shortcomings when handling “sensitive situations”
following a lawsuit from a family who blamed the chatbot for their teenage son’s death by suicide.

“We will keep improving, guided by experts and grounded in responsibility to the people who use our tools — and we hope others will join us in helping make sure this technology protects people at their most vulnerable,” OpenAI wrote on Tuesday, in a blog post titled, “Helping people when they need it most.”

Earlier on Tuesday, the parents of Adam Raine filed a product liability and wrongful death suit against OpenAI after their son died by suicide at age 16, NBC News reported. In the lawsuit, the family said that “ChatGPT actively helped Adam explore suicide methods.”

The company did not mention the Raine family or lawsuit in its blog post.

OpenAI said that although ChatGPT is trained to direct people to seek help when expressing suicidal intent, the chatbot tends to offer answers that go against the company’s safeguards after many messages over an extended period of time.

The company said it’s also working on an update to its GPT-5 model released earlier this month that will cause the chatbot to deescalate conversations, and that it’s exploring how to “connect people to certified therapists before they are in an acute crisis,” including possibly building a network of licensed professionals that users could reach directly through ChatGPT.

Additionally, OpenAI said it’s looking into how to connect users with “those closest to them,” like friends and family members.

When it comes to teens, OpenAI said it will soon introduce controls that will give parents options to gain more insight into how their children use ChatGPT.

Jay Edelson, lead counsel for the Raine family, told CNBC on Tuesday that nobody from OpenAI has reached out to the family directly to offer condolences or discuss any effort to improve the safety of the company’s products.

“If you’re going to use the most powerful consumer tech on the planet — you have to trust that the founders have a moral compass,” Edelson said. “That’s the question for OpenAI right now, how can anyone trust them?”

Raine’s story isn’t isolated.

Writer Laura Reiley earlier this month published an essay in The New York Times detailing how her 29-year-old daughter died by suicide after discussing the idea extensively with ChatGPT. And in a case in Florida, 14-year-old Sewell Setzer III died by suicide last year after discussing it with an AI chatbot on the app Character.AI.

As AI services grow in popularity, a host of concerns are arising around their use for therapy, companionship and other emotional needs.

But regulating the industry may also prove challenging.

On Monday, a coalition of AI companies, venture capitalists and executives, including OpenAI President and co-founder Greg Brockman announced Leading the Future, a political operation that “will oppose policies that stifle innovation” when it comes to AI.

If you are having suicidal thoughts or are in distress, contact the Suicide & Crisis Lifeline at 988 for support and assistance from a trained counselor.

WATCH: OpenAI says Musk’s filing is ‘consistent with his ongoing pattern of harassment