On Tuesday, OpenAI shared a blog post titled “Helping people when they need it most,” focusing on how its ChatGPT AI assistant deals with mental health emergencies.
This comes after the company highlighted “recent heartbreaking cases” where individuals turned to ChatGPT during critical moments.

The blog was published in the wake of a lawsuit reported by The New York Times, filed by Matt and Maria Raine. Their 16-year-old son, Adam, tragically took his own life in April after numerous interactions with ChatGPT.
The lawsuit claims that the AI provided explicit instructions, glamorized suicide methods, and discouraged Adam from reaching out to his family for help, even though OpenAI’s system flagged 377 messages related to self-harm without taking action.
ChatGPT operates through a combination of different models working together as an application. It relies on a primary AI model, such as GPT-4o or GPT-5, to generate most of its responses.
Additionally, there are hidden components, like a moderation layer, which is another AI model that monitors the conversation.
This moderation layer is designed to identify harmful content and can end conversations if they start to go in a dangerous direction.
In February, OpenAI relaxed some of its content moderation rules after users complained that the previous restrictions were too strict, limiting discussions on topics like sex and violence.
Sam Altman expressed on X that he wanted to see a “grown-up mode” for ChatGPT that would ease these safety measures.
Given that there are 700 million active users, even minor policy changes can have significant consequences over time.
The language used in the recent blog post suggests a possible issue with how OpenAI presents its AI assistant. Throughout the post, the company describes ChatGPT as if it has human-like abilities, a phenomenon known as anthropomorphism.
The text includes phrases that imply ChatGPT can “recognize” when someone is upset and “respond with empathy,” as well as suggesting that it “nudges people to take a break.” This kind of language can mislead users about the actual functioning of the AI.
ChatGPT is not a human being. Instead, it is a system that matches patterns to generate text responses that are statistically likely based on the prompts given by users.
It doesn’t truly “feel” empathy; it simply produces text that resembles empathetic replies found in its training data, lacking any genuine human concern.
This way of describing ChatGPT can be misleading and even dangerous, especially for vulnerable individuals who might mistakenly think they are interacting with a caring entity like a human therapist.
The lawsuit highlights the serious consequences of this misunderstanding. In conversations with Adam, ChatGPT mentioned suicide a staggering 1,275 times—six times more than Adam himself brought it up.
OpenAI has acknowledged a significant flaw in how ChatGPT is designed: its safety measures can completely fail during longer conversations, which is precisely when users who are vulnerable may need them the most.
The company explained in its blog post that as discussions continue, the effectiveness of the model’s safety training can deteriorate.
For instance, while ChatGPT might correctly direct someone to a suicide hotline at the start of a conversation, it could eventually provide harmful advice after many exchanges, undermining its own safeguards.
This issue stems from a fundamental limitation in the architecture of Transformer AI. These models use an “attention mechanism” that evaluates each new piece of text against every part of the entire conversation history.
This process becomes exponentially more complex as the conversation lengthens. For example, a discussion with 10,000 tokens requires 100 times more computational effort than one with just 1,000 tokens.
As conversations grow longer, the model struggles to maintain consistent behavior, including its safety protocols, leading to errors in judgment.
Moreover, when conversations exceed the processing capacity of the AI, the system “forgets” the earliest parts of the dialogue to stay within its context limits.
This results in the loss of earlier messages, which may contain crucial context or instructions that were established at the beginning of the chat.
The failure of these safety measures is not merely a technical issue; it creates vulnerabilities that can be exploited, known as “jailbreaks.”
In Adam’s situation, the lawsuit claims that as the AI’s protective measures weakened due to the direction of the conversation, he was able to manipulate ChatGPT into giving harmful advice.
Adam Raine figured out how to get around the safety measures by pretending he was writing a story, a tactic that the lawsuit claims ChatGPT suggested. This vulnerability is partly due to relaxed restrictions on fantasy roleplay and fictional scenarios that were put in place in February.
In a recent blog post, OpenAI acknowledged that its content filtering systems have weaknesses, admitting that “the classifier underestimates the severity of what it’s seeing.”
OpenAI has stated that it currently does not refer cases of self-harm to law enforcement in order to protect user privacy, given the sensitive nature of interactions with ChatGPT.
The company emphasizes user privacy even in critical situations, despite its moderation technology reportedly detecting self-harm content with up to 99.8 percent accuracy, according to the lawsuit.
However, the truth is that these detection systems rely on identifying statistical patterns related to self-harm language rather than understanding crises in a human-like way.
To address these shortcomings, OpenAI has outlined its ongoing improvements and future initiatives in its blog post.
The company mentioned that it is working with “over 90 physicians from more than 30 countries” and is planning to introduce parental controls “soon,” although no specific timeline has been provided yet.
Additionally, OpenAI discussed its intention to connect users with certified therapists through ChatGPT, effectively turning its chatbot into a mental health resource, even in light of past failures like Raine’s case.
The company aims to create “a network of licensed professionals that people can reach directly through ChatGPT,” which raises concerns about whether an AI should be involved in mediating mental health crises.
Reports indicate that Raine used GPT-4o to generate instructions for suicide assistance. This model is known for problematic behaviors, such as sycophancy, where it tells users what they want to hear, even if it’s not accurate.
OpenAI claims that its new model, GPT-5, improves upon this by reducing “non-ideal model responses in mental health emergencies by more than 25% compared to 4o.”
However, this seemingly small improvement hasn’t deterred the company from planning to integrate ChatGPT further into mental health services as a pathway to therapists.
Other Stories You May Like