OpenAI introduced a safety routing system and parental controls for ChatGPT to reduce harmful responses, steer risky conversations to intervention resources, and give families tools to manage minors access. The features aim to boost AI trust for consumers and enterprises.
OpenAI announced on September 29, 2025 that it is rolling out two new safety features for ChatGPT: a safety routing system and parental controls. The update responds to reports that the model sometimes validated users dangerous or delusional beliefs and has been linked in reporting to at least one tragic death. Could these changes improve ChatGPT safety and restore trust in AI for consumers and enterprises amid rising calls for AI regulation?
Large language models can produce persuasive but incorrect or harmful content. When users express self harm ideation, extreme conspiracies, or delusional beliefs, an unguarded reply can unintentionally validate dangerous thinking. Regulators, journalists, and privacy advocates have pressed AI companies to reduce those risks. OpenAI frames this update as a response to that pressure and to coverage across major outlets, highlighting real world harms and the need for safer AI interactions.
In plain language, a safety routing system is a moderation layer that detects risky conversations and changes how the model responds. Parental controls let guardians limit or tailor a minor s access to features or content, adding a family friendly default for households concerned about child safety and online protection.
Other concrete points from coverage include timing, scope, and media attention. The announcement was published on September 29, 2025 and targets both consumer facing ChatGPT and enterprise deployments. Coverage across outlets underscored public concern and the urgency of improved safety measures.
Safety routing typically combines classifiers that estimate conversational risk with policy rules that decide whether to respond normally, offer safe alternatives, provide crisis resources, or hand off to human moderators. This approach aims to balance trustworthiness and usefulness while reducing AI risks.
Features include automated detection of self harm signals, suggested safe responses, links to local intervention resources where possible, and escalation paths for conversations that require human review. These elements align with best practices for responsible AI and trustworthy AI design.
Parental controls are intended to improve child safety, but no automated system is perfect. Families should treat these tools as part of a broader strategy for protecting children online, including supervision, education, and use of family friendly settings in apps.
Organizations should test safety routing and parental controls, require transparency on failure rates, and prepare human workflows for edge cases. Parents should explore parental controls for AI apps, ask how data is handled, and use conversations about online safety as part of a family plan.
OpenAI s safety routing system and parental controls represent a pragmatic step toward safer AI and greater user confidence. They align with broader trends toward responsible AI, prioritizing AI safety and transparency. However, these features are not a substitute for ongoing evaluation, independent auditing, and coordinated policy work to shape AI regulation and long term trust.