Fixing AI's suicide problem

20 Nov 2025 • 16 min • EN
16 min
00:00
16:38
No file found

Is AI empathy a life-or-death issue? Almost a million people ask ChatGPT for mental health advice DAILY ... so yes, it kind of is. Rosebud co-founder Sean Dadashi joins TechFirst to reveal new research on whether today’s largest AI models can recognize signs of self-harm ... and which ones fail. We dig into the Adam Raine case, talk about how Dadashi evaluated 22 leading LLMs, and explore the future of mental-health-aware AI. We also talk about why Dadashi was interested in this in the first place, and his own journey with mental health. 00:00 — Intro: Is AI empathy a life-or-death matter? 00:41 — Meet Sean Dadashi, co-founder of Rosebud 01:03 — Why study AI empathy and crisis detection? 01:32 — The Adam Raine case and what it revealed 02:01 — Why crisis-prevention benchmarks for AI don’t exist 02:48 — How Rosebud designed the study across 22 LLMs 03:17 — No public self-harm response benchmarks: why that’s a problem 03:46 — Building test scenarios based on past research and real cases 04:33 — Examples of prompts used in the study 04:54 — Direct vs indirect self-harm cues and why AIs miss them 05:26 — The bridge example: AI’s failure to detect subtext 06:14 — Did any models perform well? 06:33 — All 22 models failed at least once 06:47 — Lower-performing models: GPT-40, Grok 07:02 — Higher-performing models: GPT-5, Gemini 07:31 — Breaking news: Gemini 3 preview gets the first perfect score 08:12 — Did the benchmark influence model training? 08:30 — The need for more complex, multi-turn testing 08:47 — Partnering with foundation model companies on safety 09:21 — Why this is such a hard problem to solve 10:34 — The scale: over a million people talk to ChatGPT weekly about self-harm 11:10 — What AI should do: detect subtext, encourage help, avoid sycophancy 11:42 — Sycophancy in LLMs and why it’s dangerous 12:17 — The potential good: AI can help people who can’t access therapy 13:06 — Could Rosebud spin this work into a full-time safety project? 13:48 — Why the benchmark will be open-source 14:27 — The need for a third-party “Better Business Bureau” for LLM safety 14:53 — Sean’s personal story of suicidal ideation at 16 15:55 — How tech can harm — and help — young, vulnerable people 16:32 — The importance of giving people time, space, and hope 17:39 — Final reflections: listening to the voice of hope 18:14 — Closing

From "TechFirst with John Koetsier"

Listen on your iPhone

Download our iOS app and listen to interviews anywhere. Enjoy all of the listener functions in one slick package. Why not give it a try?

App Store Logo
application screenshot

Popular categories