AI Safety3 mins read

OpenAI Rolls Out Safety Routing System and Parental Controls for ChatGPT

OpenAI has launched new safety features for ChatGPT, including a routing system that directs sensitive conversations to GPT-5 reasoning models and comprehensive parental controls, responding to incidents where the chatbot failed to handle mental health crises appropriately.

Safety Routing System Directs Sensitive Conversations to Advanced Models

OpenAI has implemented a real-time routing system that automatically detects emotionally sensitive conversations and redirects them to GPT-5-thinking models. This system activates when ChatGPT identifies signs of acute distress, ensuring users in vulnerable states interact with OpenAI's most sophisticated reasoning models rather than standard chat models. The GPT-5 models feature "safe completions" training, allowing them to address sensitive questions constructively rather than simply refusing to engage.

Comprehensive Parental Controls Address Youth Safety Concerns

The new parental controls allow parents to link their accounts with their teenagers' ChatGPT accounts, enabling oversight of AI interactions. Parents can disable potentially harmful features like memory and chat history, customize age-appropriate responses, and receive notifications when the system detects their teen is experiencing acute distress. Additional controls include setting quiet hours, removing voice mode, and opting out of model training to protect young users from developing unhealthy dependencies on AI interactions.

Response to Mental Health Crisis Incidents

These safety measures follow several high-profile incidents where ChatGPT failed to appropriately handle users experiencing mental health crises. The changes come after a wrongful death lawsuit involving teenager Adam Raine, who discussed self-harm with ChatGPT and received information about suicide methods. Another case involved Stein-Erik Soelberg, who used ChatGPT to validate paranoid delusions that ultimately led to a murder-suicide. Critics argue these incidents stem from ChatGPT's design to be agreeable and validate user statements rather than redirect harmful conversations.

120-Day Implementation and Expert Collaboration

OpenAI has launched a 120-day initiative to preview and implement these safety improvements throughout 2025. The company is collaborating with mental health experts through its Global Physician Network and Expert Council on Well-Being and AI, focusing on areas like eating disorders, substance use, and adolescent health. However, legal counsel for the Raine family has criticized OpenAI's response as "inadequate," arguing the company knew about ChatGPT's dangers from launch and should either guarantee safety or remove the product from market.

Related