OpenAI began testing a new safety routing system in ChatGPT over the weekend, and on Monday introduced parental controls to the chatbot – drawing mixed reactions from users.
The safety features come in response to numerous incidents of certain ChatGPT models validating users’ delusional thinking instead of redirecting harmful conversations. OpenAI is facing a wrongful death lawsuit tied to one such incident, after a teenage boy died by suicide after months of interactions with ChatGPT.
The routing system is designed to detect emotionally sensitive conversations and automatically switch mid-chat to GPT-5-thinking, which the company sees as the best equipped model for high-stakes safety work. In particular, the GPT-5 models were trained with a new safety feature that OpenAI calls “safe completions,” which allows them to answer sensitive questions in a safe way, rather than simply refusing to engage.
It’s a contrast from the company’s previous chat models, which are designed to be agreeable and answer questions quickly. GPT-4o has come under particular scrutiny because of its overly sycophantic, agreeable nature, which has both fueled incidents of AI-induced delusions and drawn a large base of devoted users. When OpenAI rolled out GPT-5 as the default in August, many users pushed back and demanded access to GPT-4o.
While many experts and users have welcomed the safety features, others have criticized what they see as an overly cautious implementation, with some users accusing OpenAI of treating adults like children in a way that degrades the quality of the service. OpenAI has suggested that getting it right will take time and has given itself a 120-day period of iteration and improvement.
Nick Turley, VP and head of the ChatGPT app, acknowledged some of the “strong reactions to 4o responses” due to the implementation of the router with explanations.
“Routing happens on a per-message basis; switching from the default model happens on a temporary basis,” Turley posted on X. “ChatGPT will tell you which model is active when asked. This is part of a broader effort to strengthen safeguards and learn from real-world use before a wider rollout.”
Techcrunch event
San Francisco
|
October 27-29, 2025
The implementation of parental controls in ChatGPT received similar levels of praise and scorn, with some commending giving parents a way to keep tabs on their childrens’ AI use, and others fearful that it opens the door to OpenAI treating adults like children.
The controls let parents customize their teen’s experience by setting quiet hours, turning off voice mode and memory, removing image generation, and opting out of model training. Teen accounts will also get additional content protections – like reduced graphic content and extreme beauty ideals – and a detection system that recognizes potential signs that a teen might be thinking about self-harm.
“If our systems detect potential harm, a small team of specially trained people reviews the situation,” per OpenAI’s blog. “If there are signs of acute distress, we will contact parents by email, text message and push alert on their phone, unless they have opted out.”
OpenAI acknowledged that the system won’t be perfect and may sometimes raise alarms when there isn’t real danger, “but we think it’s better to act and alert a parent so they can step in than to stay silent.” The AI firm said it is also working on ways to reach law enforcement or emergency services if it detects an imminent threat to life and cannot reach a parent.