← Blog

AI tools don't have a hallucination problem. They have an amnesia problem.

Vitaliy Filipov · · 6 min read

Hallucination gets all the attention. Fair enough. A confidently wrong answer can ruin your afternoon. But I think we’re focused on the wrong problem.

Your AI tools forget everything, every single session.

Last week I spent 20 minutes getting Claude to understand our authentication flow. Got a solid refactor out of it. Opened a new session the next morning, and it had no idea what OAuth2 provider we use, what the middleware looks like, or that we migrated to a new pattern on Tuesday.

That’s not hallucination. That’s amnesia. And while half the industry is building hallucination guardrails, almost nobody is working on this.

The forgetting tax

Every re-explanation has a cost. I’ve started calling it the forgetting tax, the cumulative time your team burns re-teaching AI tools things they already knew.

Back-of-napkin math:

Per sessionPer engineer/weekTeam of 8/month
Context setup15-30 min1.5-3 hours48-96 hours
Stale answers corrected5-10 min30-60 min16-32 hours
Insights lostunmeasurablecompoundscompounds

A team of eight engineers can lose 60-120 hours per month to re-explanations. That’s roughly a full-time engineer’s output. Gone.

But the table understates it. The real damage is what you can’t count: the fix someone found on Tuesday that would’ve prevented Thursday’s incident, except the session ended and the knowledge vanished.

Why this keeps happening

The problem isn’t broken software. It’s a missing layer. AI tools today have three types of memory, and most teams are only using one:

graph TD
  T1["Tier 1: Session Memory"] --- D1["Lives in the context window"]
  T2["Tier 2: Persistent Memory"] --- D2["Remembers per user across sessions"]
  T3["Tier 3: Shared Memory"] --- D3["Team-wide, always current, tool-agnostic"]
  T1 -.- L1["Claude, Cursor, ChatGPT"]
  T2 -.- L2["ChatGPT Memory, Mem0"]
  T3 -.- L3["Almost nobody"]
  style T1 fill:#1e1e3a,stroke:#6366f1,color:#e2e8f0
  style T2 fill:#1e1e3a,stroke:#6366f1,color:#e2e8f0
  style T3 fill:#1e1e3a,stroke:#22c55e,color:#e2e8f0
  style L3 fill:#0f172a,stroke:#22c55e,color:#22c55e
Click to expand
Three tiers of AI memory: most tools only implement Tier 1

Tier 1 (Session memory) is what every AI tool has today. It’s the context window. Powerful while the session is open, gone the moment you close the tab.

Tier 2 (Persistent memory) is what a few tools are starting to offer. ChatGPT remembers your preferences. Mem0 stores facts per user. Zep adds temporal reasoning. Better than nothing, but it’s still your memory, not the team’s. (We reviewed all the major tools in our AI Memory Landscape.)

Tier 3 (Shared memory) is the missing layer. When one engineer figures out the fix, everyone’s AI tools learn it too. The migration happened last Tuesday? The memory layer watched it happen and updated itself. Nobody filed a ticket.

Four ways amnesia breaks your team

Wasted time is just the surface. Amnesia creates four failure modes that compound on each other:

graph TD
  A["AI Amnesia"] --> P["Stale Confidence"]
  A --> D["Context Silos"]
  A --> C["Lost Solutions"]
  A --> X["Ghost Knowledge"]
  P --- PD["AI gives outdated answers with full confidence"]
  D --- DD["Cursor knows things Claude doesn't and vice versa"]
  C --- CD["Tuesday's fix is forgotten by Thursday"]
  X --- XD["Knowledge exists somewhere but nobody can find it"]
  style A fill:#1e1e3a,stroke:#ef4444,color:#e2e8f0
  style P fill:#1e1e3a,stroke:#f59e0b,color:#e2e8f0
  style D fill:#1e1e3a,stroke:#f59e0b,color:#e2e8f0
  style C fill:#1e1e3a,stroke:#f59e0b,color:#e2e8f0
  style X fill:#1e1e3a,stroke:#f59e0b,color:#e2e8f0
Click to expand
The four failure modes of AI amnesia

Stale confidence

Your team deprecated a library last week. The migration doc exists, but the old getting-started guide hasn’t been updated yet. An AI tool retrieves the old guide because it’s a closer semantic match to the question, and confidently recommends the deprecated pattern.

This is arguably worse than not having AI at all. Without AI, the engineer would just ask a teammate and get the right answer.

Context silos

Engineer A spends an hour teaching Cursor about the payment service architecture. Engineer B opens Claude and asks the same question. Claude has no idea. The knowledge exists, but it’s locked in A’s session.

Congratulations, you now have five knowledge silos instead of one. Docs, Claude, Cursor, ChatGPT, and whatever custom agent someone built last sprint.

Lost solutions

A senior engineer debugs a gnarly race condition in the queue consumer. The session captures the full reasoning chain: the hypothesis, the investigation, the root cause, the fix. Session ends. It’s all gone.

Three months later, someone hits the same bug. No one remembers the fix. The investigation starts from scratch.

Ghost knowledge

This one is maddening. The answer exists, buried in an old session, in someone’s chat history, in a doc that links to a doc that references a Slack thread from three months ago. But finding it takes longer than figuring it out from scratch. So you figure it out from scratch. Again.

What the fix looks like

Bigger context windows won’t fix this. Neither will better RAG. The fix is a memory layer that operates at the team level, not per user, not per tool:

graph LR
  E1["Engineer A in Cursor"] -->|MCP| M["Shared Memory Layer"]
  E2["Engineer B in Claude"] -->|MCP| M
  E3["CI/CD Pipeline"] -->|API| M
  E4["Custom Agent"] -->|MCP| M
  M --> K1["Architectural Decisions"]
  M --> K2["Code Patterns"]
  M --> K3["Incident History"]
  M --> K4["Team Conventions"]
  style M fill:#1e1e3a,stroke:#22c55e,color:#e2e8f0
Click to expand
How shared memory connects your team's AI tools

What does this actually need to do?

Cursor, Claude, and your custom agents should all read from and write to the same knowledge base. When code changes or decisions are made, the memory should reflect it without anyone having to “update the docs.” Every fact needs a source, an author, and a timestamp. And workspaces need isolation so teams only see what they should.

The point: memory isn’t a feature you bolt onto individual tools. It’s infrastructure, same way a database sits beneath your application services.

The compounding effect

The thing that gets me excited about shared memory isn’t the time savings. It’s that it compounds.

Day one, you save 20 minutes of context setup. Fine. Week one, the new hire’s AI tools already know the codebase conventions without anyone explaining them. Month one, every bug fix, architecture decision, and code review feeds back into the collective knowledge. By month three, your AI tools understand your systems better than any individual engineer does, because they’ve absorbed everyone’s contributions.

The forgetting tax compounds in the wrong direction. Shared memory compounds in the right one.

Getting started

You don’t need to replace anything. MCP (Model Context Protocol) lets you plug a memory layer into whatever AI tools your team already uses: Claude, Cursor, custom agents, all of it. One config change per tool.

The question isn’t whether your team needs this. It’s how many more hours of re-explanations you’re willing to sit through before you set it up.