Your AI Therapist Forgets You Every Time You Log In. Here's Why That's a Problem.
- Udayaditya Barua

- 7 days ago
- 2 min read
Standard AI chatbots forget you every session. Learn why stateless LLMs fail at mental health support and how longitudinal memory changes everything.
Imagine walking into your therapist's office. You sit down, ready to pick up where you left off last week about your work anxiety.
Your therapist looks at you blankly and asks: "Who are you, and why are you here?"
You spend the first 20 minutes re-explaining your name, your job, your childhood, your trauma. Finally, you get 10 minutes of actual advice. Then you leave.
Next week, you come back. Your therapist stares at you blankly again. "I'm sorry, who are you?"
This would be malpractice in therapy. Yet this is exactly how most people interact with ChatGPT, Claude, and Gemini every single day.
The "Stateless" Problem
In engineering terms, standard LLMs (Large Language Models) are stateless.
Every time you close the tab or start a new chat, the "state" resets to zero. The model has no idea who you are. It has no object permanence. To the AI, you are a stranger every time you open the app.
This creates two massive problems for anyone using AI for mental health support:
1. Onboarding Fatigue - You have to re-explain your context every single time you want advice. "I'm a founder." "I have anxiety about deadlines." "My father was critical growing up." It's exhausting, and it stops people from actually using the tool when they need it most.
2. Recency Bias - Because the AI only knows what you just typed in the current session, it gives shallow advice. It can't see the pattern forming over six months—it can only see the text from the last five minutes. It's like trying to diagnose someone's health from a single snapshot instead of their medical history.
Building "Stateful" AI
When we architected Flammingo, we knew we couldn't just build another API wrapper. We had to build a memory layer.
We moved from a stateless architecture (New Chat = New Person) to a stateful one (New Chat = Same Person, New Context).
We achieve this using what we call Longitudinal Memory.
Instead of feeding the AI just your current message, our backend retrieves relevant "shards" of your past conversations to give the AI context before it even responds.
Here's what that looks like in practice:
You type: "I feel overwhelmed today."
Stateless AI thinks: "User is stressed. Give generic stress advice."
Flammingo thinks: "User is overwhelmed. Retrieving history... Context found: User reports feeling overwhelmed every Tuesday before investor updates. Action: Ask if today's investor update is the trigger."
The difference isn't just technical - it's the difference between talking to a stranger and talking to someone who knows you.
The Death of the Context Window
The tech industry is obsessed with "context windows"—how much text you can paste into a single prompt. Companies compete over whether they can handle 100K tokens or 200K tokens.
But a mental health journey isn't a window. It's a timeline.
By mapping your history into a vector database, we don't just "remember" what you said—we understand when you said it, how often it happens, and what usually helps.
We killed the reset button.
Because you shouldn't have to introduce yourself to your own journal.

Comments