Is AI empathy a life-or-death issue? Almost a million people ask ChatGPT for mental health advice DAILY ... so yes, it kind of is. Rosebud co-founder Sean Dadashi joins TechFirst to reveal new research on whether today’s largest AI models can recognize signs of self-harm ... and which ones fail. We dig into the Adam Raine case, talk about how Dadashi evaluated 22 leading LLMs, and explore the future of mental-health-aware AI. We also talk about why Dadashi was interested in this in the first place, and his own journey with mental health. 00:00 — Intro: Is AI empathy a life-or-death matter? 00:41 — Meet Sean Dadashi, co-founder of Rosebud 01:03 — Why study AI empathy and crisis detection? 01:32 — The Adam Raine case and what it revealed 02:01 — Why crisis-prevention benchmarks for AI don’t exist 02:48 — How Rosebud designed the study across 22 LLMs 03:17 — No public self-harm response benchmarks: why that’s a problem 03:46 — Building test scenarios based on past research and real cases 04:33 — Examples of prompts used in the study 04:54 — Direct vs indirect self-harm cues and why AIs miss them 05:26 — The bridge example: AI’s failure to detect subtext 06:14 — Did any models perform well? 06:33 — All 22 models failed at least once 06:47 — Lower-performing models: GPT-40, Grok 07:02 — Higher-performing models: GPT-5, Gemini 07:31 — Breaking news: Gemini 3 preview gets the first perfect score 08:12 — Did the benchmark influence model training? 08:30 — The need for more complex, multi-turn testing 08:47 — Partnering with foundation model companies on safety 09:21 — Why this is such a hard problem to solve 10:34 — The scale: over a million people talk to ChatGPT weekly about self-harm 11:10 — What AI should do: detect subtext, encourage help, avoid sycophancy 11:42 — Sycophancy in LLMs and why it’s dangerous 12:17 — The potential good: AI can help people who can’t access therapy 13:06 — Could Rosebud spin this work into a full-time safety project? 13:48 — Why the benchmark will be open-source 14:27 — The need for a third-party “Better Business Bureau” for LLM safety 14:53 — Sean’s personal story of suicidal ideation at 16 15:55 — How tech can harm — and help — young, vulnerable people 16:32 — The importance of giving people time, space, and hope 17:39 — Final reflections: listening to the voice of hope 18:14 — Closing
From "TechFirst with John Koetsier"
Comments
Add comment Feedback