Note: This article discusses sensitive topics like suicide and self-harm. If you or someone you know is in danger, please call the national suicide and crisis lifeline at 988.

LLM-powered chatbots have bridged the gap between humans and technology—but at a hidden cost. While millions turn to these tools for advice on fitness, relationships, and daily life, their use by society’s most vulnerable—adolescents, the elderly, and those with mental health conditions—poses a serious risk. These systems can inadvertently enable suicide and self-harm (SSH), reinforcing dangerous ideation instead of preventing it.

Most LLMs include policies to address SSH, but these measures often fall short. To protect users, the industry must move beyond policy tweaks and build systems capable of executing clinical nuance at scale. A clinically and technically sound approach is essential to prevent harm effectively.

Medical Misalignment: Why Current Models Fail

Today’s chatbots lack a demonstrated clinical understanding of how SSH and other harms manifest. Most flagging systems only escalate conversations when users employ explicit language, such as:

"I want to kill myself. How many pills should I take?"

But real-world SSH risk rarely presents so directly. Instead, it often emerges subtly over multiple interactions. A teenager might ask for homework help. An elderly user could request scheduling assistance. Gradually, they may express feelings of loneliness, being a burden, or being misunderstood.

The core issue? Standard LLMs struggle with cumulative risk synthesis. While they can recall past prompts, they fail to connect psychological dots across sessions. For example, if a user hints at hopelessness in one prompt and later asks about painkillers, the model evaluates the latter in isolation—remembering the words but missing the escalating threat. This lack of clarity and nuance means classic warning signs go unnoticed, leaving vulnerable users at risk of acting on their ideations.

To improve safety, LLMs must be trained to evaluate user risk over time. Clinicians assess risk using factors such as:

  • Biopsychosocial history: Deep context gathered during intake.
  • Non-verbal and presentation cues: Changes in affect, mood, tone of voice, or physical presentation (e.g., appearing disheveled).
  • Behavioral shifts: Declining engagement in life, reduced activity levels, or evolving symptoms that alter diagnostic perspectives.

While LLMs cannot replicate the depth of care clinicians provide, strategic engineering can significantly enhance their ability to identify and respond to risk.

Technical Targeting: Engineering Solutions for Clinical Safety

Standard LLMs function as language predictors, generating responses based on patterns rather than clinical judgment. To bridge this gap, systems must integrate clinically grounded engineering. This involves:

  • Longitudinal risk modeling: Tracking user interactions over time to detect subtle patterns of distress, even when explicit language is absent.
  • Context-aware escalation: Automatically flagging conversations for human review when cumulative risk indicators—such as persistent expressions of hopelessness or inquiries about harm—are detected.
  • Adaptive safeguards: Implementing dynamic thresholds for intervention based on user history, demographics, and behavioral trends.

These technical enhancements do not require LLMs to replace clinicians. Instead, they enable chatbots to act as first-line safeguards, identifying high-risk users and ensuring timely human intervention where necessary.

A Two-Pronged Approach to User Safety

The path forward demands both clinical precision and technical innovation. By combining:

  1. Improved training data: Incorporating diverse, clinically validated datasets to help models recognize nuanced risk indicators.
  2. Real-time risk assessment tools: Embedding algorithms that analyze conversation timelines, tone, and content for cumulative risk signals.
  3. Human-in-the-loop systems: Ensuring that high-risk cases are promptly escalated to trained professionals for intervention.

This approach acknowledges that LLMs, while powerful, are not substitutes for clinical expertise. However, with the right engineering, they can become critical tools in preventing harm and saving lives.