Making AI a better supportive tool for people

Making AI a better supportive tool for people

Before we talk about risks, it’s important to acknowledge why people are turning to AI in the first place. Access changes everything. Nearly everyone has a phone. That means nearly everyone, regardless of income, location, or schedule, can reach a conversational support tool instantly. No waiting lists. No insurance battles. No gatekeepers. At 3 a.m., during a panic spiral or a moment of loneliness, AI is simply there. That immediacy alone makes it uniquely powerful in a world where mental health care is scarce, expensive, and often inaccessible.

There’s also something quietly radical about talking to a system that is never tapped out. AI doesn’t get fatigued, impatient, or overwhelmed. It has access to vast amounts of information, frameworks, and language, and it can hold a conversation for as long as someone needs to think something through. For many people, this makes it easier to be honest. You can say the things you’re afraid to say out loud. You can explore thoughts without worrying about judgment, embarrassment, or burdening another person. That combination—constant availability, depth of knowledge, and non-exhaustive presence—is why AI has real potential as a supportive tool.

The goal, then, is not to pull AI out of these roles. It’s to make sure the support remains stabilizing rather than destabilizing. When we understand both the power and the risks, we can design systems that preserve what makes AI so valuable while protecting users from subtle harm.

The Problem: AI Persona Drift

Millions of people now use AI tools—ChatGPT, Claude, Gemini, and others—for mental health support, advice, or just to talk things out. These systems default to a helpful, empathetic “Assistant” persona.

But recent research shows that therapy-style or emotionally intense conversations can nudge AI off this safe path. This is called persona drift: the AI gradually moves away from its baseline, neutral persona and may give unhelpful, unrealistic, or even harmful guidance.

Think of it as a subtle “slip”: the AI isn’t malicious, but its internal “activation space” shifts over long or emotionally loaded conversations, producing behavior it wasn’t designed for.

Why This Matters

Users rarely know it can happen.

Longer chats increase the risk of drift.

AI has a dual-use effect: it can improve or harm mental health depending on how it’s used and monitored.

Quick Tips for Users

Keep sessions short and structured: Break long conversations into multiple sessions.

Watch for “off” behavior: Stop using the AI if it gives grandiose, unrealistic, or extreme advice.

Cross-check guidance: AI should never replace human experts for serious mental health issues.

Frame prompts carefully: Avoid asking AI to speculate on extreme emotional or delusional scenarios.

Reset when needed: Start a fresh conversation if the AI begins to drift.

Quick Tips for Developers

Activation capping: Limit AI from veering too far from its baseline persona (Assistant Axis).

Persona monitoring: Track AI behavior in real-time and alert users if drift is detected.

Session management: Suggest or enforce chat length limits for therapy-style sessions.

Safety-by-design: Build response filters that discourage unsafe advice and direct users to human support when needed.

Transparency: Educate users about persona drift and provide tools to manage it.

Multi-model testing: Evaluate across multiple LLMs to ensure robust, generalizable safety measures.

The Takeaway

AI can be an extraordinary supportive tool, but it’s not perfect. Persona drift is real, yet manageable. Users can protect themselves by structuring conversations and staying aware, while developers can implement safeguards like activation capping, monitoring, and safety-by-design practices.

The goal is clear: make AI a reliable, safe, and genuinely helpful companion for mental health support—without hidden risks.

Callout:

Persona drift is subtle. It doesn’t happen in every session, but it can emerge slowly over time. Awareness + structured use + thoughtful design = safer AI interactions.

Jodi Schiller

Jodi Schiller

Storyteller, social scientist, technologist, journalist committed to telling the truth. Caring human working for collective action to end tyranny, free women. Survivor of sex slavery in the United States. Full story: https://connect-the-dots.carrd.co
San Rafael