AI tools don't have a hallucination problem. They have an amnesia problem.
Hallucination gets all the attention. Fair enough. A confidently wrong answer can ruin your afternoon. But I think we’re focused on the wrong problem.
Your AI tools forget everything, every single session.
Last week I spent 20 minutes getting Claude to understand our authentication flow. Got a solid refactor out of it. Opened a new session the next morning, and it had no idea what OAuth2 provider we use, what the middleware looks like, or that we migrated to a new pattern on Tuesday.
That’s not hallucination. That’s amnesia. And while half the industry is building hallucination guardrails, almost nobody is working on this.
The forgetting tax
Every re-explanation has a cost. I’ve started calling it the forgetting tax, the cumulative time your team burns re-teaching AI tools things they already knew.
Back-of-napkin math:
| Per session | Per engineer/week | Team of 8/month | |
|---|---|---|---|
| Context setup | 15-30 min | 1.5-3 hours | 48-96 hours |
| Stale answers corrected | 5-10 min | 30-60 min | 16-32 hours |
| Insights lost | unmeasurable | compounds | compounds |
A team of eight engineers can lose 60-120 hours per month to re-explanations. That’s roughly a full-time engineer’s output. Gone.
But the table understates it. The real damage is what you can’t count: the fix someone found on Tuesday that would’ve prevented Thursday’s incident, except the session ended and the knowledge vanished.
Why this keeps happening
The problem isn’t broken software. It’s a missing layer. AI tools today have three types of memory, and most teams are only using one:
graph TD T1["Tier 1: Session Memory"] --- D1["Lives in the context window"] T2["Tier 2: Persistent Memory"] --- D2["Remembers per user across sessions"] T3["Tier 3: Shared Memory"] --- D3["Team-wide, always current, tool-agnostic"] T1 -.- L1["Claude, Cursor, ChatGPT"] T2 -.- L2["ChatGPT Memory, Mem0"] T3 -.- L3["Almost nobody"] style T1 fill:#1e1e3a,stroke:#6366f1,color:#e2e8f0 style T2 fill:#1e1e3a,stroke:#6366f1,color:#e2e8f0 style T3 fill:#1e1e3a,stroke:#22c55e,color:#e2e8f0 style L3 fill:#0f172a,stroke:#22c55e,color:#22c55eClick to expand
Tier 1 (Session memory) is what every AI tool has today. It’s the context window. Powerful while the session is open, gone the moment you close the tab.
Tier 2 (Persistent memory) is what a few tools are starting to offer. ChatGPT remembers your preferences. Mem0 stores facts per user. Zep adds temporal reasoning. Better than nothing, but it’s still your memory, not the team’s. (We reviewed all the major tools in our AI Memory Landscape.)
Tier 3 (Shared memory) is the missing layer. When one engineer figures out the fix, everyone’s AI tools learn it too. The migration happened last Tuesday? The memory layer watched it happen and updated itself. Nobody filed a ticket.
Four ways amnesia breaks your team
Wasted time is just the surface. Amnesia creates four failure modes that compound on each other:
graph TD A["AI Amnesia"] --> P["Stale Confidence"] A --> D["Context Silos"] A --> C["Lost Solutions"] A --> X["Ghost Knowledge"] P --- PD["AI gives outdated answers with full confidence"] D --- DD["Cursor knows things Claude doesn't and vice versa"] C --- CD["Tuesday's fix is forgotten by Thursday"] X --- XD["Knowledge exists somewhere but nobody can find it"] style A fill:#1e1e3a,stroke:#ef4444,color:#e2e8f0 style P fill:#1e1e3a,stroke:#f59e0b,color:#e2e8f0 style D fill:#1e1e3a,stroke:#f59e0b,color:#e2e8f0 style C fill:#1e1e3a,stroke:#f59e0b,color:#e2e8f0 style X fill:#1e1e3a,stroke:#f59e0b,color:#e2e8f0Click to expand
Stale confidence
Your team deprecated a library last week. The migration doc exists, but the old getting-started guide hasn’t been updated yet. An AI tool retrieves the old guide because it’s a closer semantic match to the question, and confidently recommends the deprecated pattern.
This is arguably worse than not having AI at all. Without AI, the engineer would just ask a teammate and get the right answer.
Context silos
Engineer A spends an hour teaching Cursor about the payment service architecture. Engineer B opens Claude and asks the same question. Claude has no idea. The knowledge exists, but it’s locked in A’s session.
Congratulations, you now have five knowledge silos instead of one. Docs, Claude, Cursor, ChatGPT, and whatever custom agent someone built last sprint.
Lost solutions
A senior engineer debugs a gnarly race condition in the queue consumer. The session captures the full reasoning chain: the hypothesis, the investigation, the root cause, the fix. Session ends. It’s all gone.
Three months later, someone hits the same bug. No one remembers the fix. The investigation starts from scratch.
Ghost knowledge
This one is maddening. The answer exists, buried in an old session, in someone’s chat history, in a doc that links to a doc that references a Slack thread from three months ago. But finding it takes longer than figuring it out from scratch. So you figure it out from scratch. Again.
What the fix looks like
Bigger context windows won’t fix this. Neither will better RAG. The fix is a memory layer that operates at the team level, not per user, not per tool:
graph LR E1["Engineer A in Cursor"] -->|MCP| M["Shared Memory Layer"] E2["Engineer B in Claude"] -->|MCP| M E3["CI/CD Pipeline"] -->|API| M E4["Custom Agent"] -->|MCP| M M --> K1["Architectural Decisions"] M --> K2["Code Patterns"] M --> K3["Incident History"] M --> K4["Team Conventions"] style M fill:#1e1e3a,stroke:#22c55e,color:#e2e8f0Click to expand
What does this actually need to do?
Cursor, Claude, and your custom agents should all read from and write to the same knowledge base. When code changes or decisions are made, the memory should reflect it without anyone having to “update the docs.” Every fact needs a source, an author, and a timestamp. And workspaces need isolation so teams only see what they should.
The point: memory isn’t a feature you bolt onto individual tools. It’s infrastructure, same way a database sits beneath your application services.
The compounding effect
The thing that gets me excited about shared memory isn’t the time savings. It’s that it compounds.
Day one, you save 20 minutes of context setup. Fine. Week one, the new hire’s AI tools already know the codebase conventions without anyone explaining them. Month one, every bug fix, architecture decision, and code review feeds back into the collective knowledge. By month three, your AI tools understand your systems better than any individual engineer does, because they’ve absorbed everyone’s contributions.
The forgetting tax compounds in the wrong direction. Shared memory compounds in the right one.
Getting started
You don’t need to replace anything. MCP (Model Context Protocol) lets you plug a memory layer into whatever AI tools your team already uses: Claude, Cursor, custom agents, all of it. One config change per tool.
The question isn’t whether your team needs this. It’s how many more hours of re-explanations you’re willing to sit through before you set it up.
Venture Builder @ Knowledge Plane
Obsessed with high-performing teams and validation-driven development. Forbes 30 under 30. Building at Camplight and Knowledge Plane.