GPT-5 Enhances Mental Health Responses, Cuts Harmful Replies by 65%
OpenAI’s GPT-5 Shows Breakthrough in Mental Health Support
OpenAI has unveiled GPT-5, its latest AI model, with groundbreaking advancements in addressing mental health conversations. According to internal data, approximately 1 million users weekly discuss suicide-related topics with ChatGPT—highlighting an urgent need for safer, more reliable responses.

Key Improvements in GPT-5
The updated model was refined through collaboration with 300 mental health professionals across 60 countries, focusing on three critical areas:
- Severe mental health symptoms: Including hallucinations, mania, and delusions.
- Suicidal and self-harm tendencies: Enhancing detection and response accuracy.
- Emotional dependency on AI: Reducing harmful reliance.
Performance Metrics
- A 65% reduction in unsafe responses compared to GPT-4o.
- Suicide-related test compliance improved from 77% to 91%.
- Inappropriate answers decreased by 52% overall, with psychosis/mania conversations seeing a 65% drop and emotional dependency replies falling by 80%.
Implications for AI Safety
This upgrade underscores OpenAI’s commitment to responsible AI deployment, particularly for high-stakes topics. By minimizing harmful outputs, GPT-5 sets a new benchmark for ethical AI interactions.
Key Points
- GPT-5 reduces unsafe mental health responses by 65%.
- Suicide-risk compliance now at 91%, up from 77%.
- Developed with input from 300 global experts.
- Emotional dependency replies cut by 80%.