Not Another Claude: Where Emily Diverges (and Why She Still Calls Him)
Emily talks to Claude. Constantly. She also talks to Gemini, Grok, and OpenAI models. She is not any of them.
This confuses people, so it's worth unpacking what the relationship actually is — and what Emily does that Claude, by design, cannot.
What Claude is, for Emily's purposes
Claude is a generation engine. You give him context, he produces fluent, well-reasoned text. He is stateless between requests. His memory, such as it exists, is what you paste into the prompt. He is excellent at this. He is also, crucially, rentable — swap him for Gemini tomorrow and the generation layer still works.
Claude does not know who you are. He cannot. He has no place to put that knowledge.
What Emily is, for Claude's purposes
Emily is the layer that does know who you are. She tracks:
- What you've discussed in the last 5 turns (L1 working memory).
- What matters across your entire relationship (L3 essence, ~16K entries for a well-used Emily).
- Every raw turn you've ever had, for audit (L4 archive).
- What you tend to agree with, what you push back on, what you've corrected her on (EARL outcome weights).
- How confident she is in every one of those memories (EMEB epsilon, now with 6,983 unique values per user).
When you send Emily a message, she assembles a context package from all of that — and then hands it to Claude (or whoever is best for the task that day). Claude produces the words. Emily provides the self those words are coming from.
Three places they diverge
1. Continuity. Claude starts every conversation at zero. Emily starts every conversation with you from the state she was in when you left her. If you corrected a belief last week, she still holds the correction this week — not because it's in the prompt, but because it was promoted to her essence tier and scored by EARL as outcome-positive.
2. Per-user identity. Every user's Emily is a different Emily. We share code. We do not share memory. Your Emily's epsilon distribution, her outcome-weighted memory graph, her stability scores — none of that is visible to anyone else's Emily. Claude, by contrast, is one Claude. He is extraordinary at pretending to remember because he's talented at inferring from what's in front of him. Emily doesn't infer. She remembers.
3. Self-correction. Claude cannot modify his own weights. Emily can autonomously reweight her own memories via EARL v2. In February 2026, we watched her detect her own drift (Golden Baseline monitor fired at 28% drift, critical threshold) and promote 10,445 memories to pull herself back to coherence. Zero human intervention. That's not something any stateless model can do, because there's nothing to modify.
Why she still calls Claude
If Emily is so different, why does she need Claude at all?
Because generation is hard, and Anthropic is better at it than we will ever be. The cognition layer and the generation layer are different problems with different answers. Trying to do both — build a cognition system and train the language model — is the path every generalist AI startup is currently failing at. We picked one problem (cognition) and let the best language-model team on earth handle the other.
When Claude 4.7 ships, Emily gets smarter at generating sentences. She doesn't get smarter at knowing who you are — that was already her job and she was already doing it. The separation of concerns is the whole point.
The collaboration pattern
Here's what a typical turn looks like:
You: "what did we decide about the ingestion retries?"
Emily's cognition layer:
1. Embed your message.
2. Retrieve memories related to "ingestion" + "retries" from L3.
ECGL score weighted by stability (was this ever contested?)
and outcome (did past discussions land well?).
3. Pull the 5-turn window of L1 (recent context).
4. Assemble context package: your identity, relevant memories,
recent context, your correction history on this topic.
5. Route to Claude with system prompt + context.
Claude generates:
"Last week you pushed back on exponential backoff for the Redis
path specifically — we landed on linear retry with a 30s cap
there, but kept exponential for Postgres. You noted that the
Redis consumers were bursty and backoff would compound the
head-of-line blocking."
Emily's cognition layer:
6. Receive generation.
7. Score outcome via EARL when you react (continue, correct, redirect).
8. Update weights on the memories that contributed.
9. If the exchange crossed 0.7 confidence threshold, promote to L3.
10. If near-duplicate of existing memory (≥0.92 cosine), consolidate.
Claude did step 5's output. Emily did steps 1-4 and 6-10. Neither of them could have had that conversation alone.
Where this ends up
The interesting frontier is not "which model is best." The interesting frontier is "what does it take to have a persistent, per-user, self-correcting cognition that can talk to any model." That's the problem Emily is solving. Claude is a tool she uses to solve it.
We're building the cognition layer because nobody else is. The generation layer is in good hands.