You've just spent forty-five minutes explaining your system architecture to an AI assistant. The service boundaries, the database migration history, why that one endpoint uses a different auth pattern. The decisions behind the decisions. It understood. It helped you solve the problem beautifully.

The next morning, you open a new session.

It doesn't remember any of it.

So you start explaining again. But this time, you skip a few details — you're tired of repeating yourself. The AI's suggestions are worse because it's working from thinner context. You correct it twice, then give up and just write the code yourself.

If this sounds familiar, you're paying the rediscovery tax.

The Tax Nobody Talks About

We talk about AI productivity in terms of what it can generate: lines of code, documents, analyses. But we rarely talk about what it forgets. And the cost of that forgetting isn't just the time spent re-explaining — it's everything that compounds from there.

The explanations get shorter because you're tired of repeating them. The AI's output gets worse because it's working from incomplete context. Architecture decisions get rediscovered instead of built upon. Mistakes that were debugged last week get made again. Style preferences that were calibrated over hours revert to defaults.

Knowledge doesn't accumulate. Every session is a cold start.

Andrej Karpathy compared it to anterograde amnesia — the condition depicted in Memento. Your AI collaborator can function brilliantly within a conversation. But the moment that conversation ends, the slate is wiped clean. It's not that the AI is incapable. It's that it has no yesterday.

Quantifying the Damage

The developer community has started calling this the "Groundhog Day" problem — and the data is sobering.

12 min
average daily re-orientation time per developer
Practitioner tracking study, 2025
19%
slower — experienced developers using AI on real codebases
METR, 2025 (16 devs, 246 tasks)
22–44%
of project budget consumed by "confusion taxes" in one agent system
Multi-agent workflow analysis, 2026
95%
of enterprise AI pilots with zero measurable P&L impact
MIT NANDA, 2025

That last number deserves a pause. Enterprises spent an estimated $644 billion on AI in 2025. Ninety-five percent of pilot projects produced zero measurable impact on the bottom line. Many analyses tie the failures explicitly to what they call "context drift" — the AI's inability to learn from previous interactions and retain institutional knowledge.

The METR study is particularly striking because it measured experienced developers working on real codebases — large, mature open-source projects with over a million lines of code. The developers perceived a 20–24% speedup. The objective measurement showed a 19% slowdown. The acceptance rate for AI-generated code was below 44%. The gap between perceived and actual productivity was 39 percentage points.

Why? Because the AI lacked tacit knowledge — the unwritten understanding of why the code is structured the way it is. Every session, it had to infer that context from scratch. And its inferences were often wrong in ways that required careful review to catch.

Why This Is Getting Worse, Not Better

A common response is: "Context windows are getting bigger. The problem will solve itself." This is a comforting thought. It's also wrong.

Larger context windows help within a session. They don't help across sessions. You can fit more documents into a single conversation, but the moment that conversation ends, every insight, every calibration, every hard-won understanding vanishes.

And there's a subtler problem. Research from Chroma and multiple academic groups has documented what's called "context rot" — even within a long session, accuracy degrades sharply as the context grows. Facts mentioned early in a conversation are progressively lost or distorted. One study showed accuracy dropping from 77% to 26% as context scaled. Organized project documentation sometimes performed worse than random order.

Bigger windows are a larger short-term memory. They are not long-term understanding. The analogy is precise: giving someone a larger desk doesn't give them a better memory. It gives them more space to spread out papers they'll sweep into the bin at the end of the day.

The Compounding Problem

Simple tasks — "write a function that sorts this list" — don't suffer much from AI amnesia. The context fits in the prompt. But the value of AI collaboration scales with complexity. And complexity is exactly where amnesia hurts most.

A ten-session project doesn't lose 10× the context of a one-session task. It loses the relationships between sessions — the evolving understanding, the dead ends that were eliminated, the design decisions that were debated and resolved. The compound interest of knowledge. That's the real tax.

Retrieval Is Not Memory

The market's current answer to AI amnesia is retrieval-augmented generation — RAG. Embed your documents, search for relevant chunks, inject them into the prompt. It's become the default architecture. It's also insufficient.

RAG gives AI access to information. It does not give it understanding. The difference matters.

Memory — the kind that makes collaboration compound — requires things that retrieval doesn't provide:

The AI memory ecosystem is maturing rapidly — Mem0, Zep, Letta, and others are building increasingly sophisticated approaches. But the problem space is harder than it appears. Extraction noise, stale facts, conflict resolution, and the governance of persistent knowledge are all unsolved at scale. The field is moving from "store everything" to "understand what matters" — and that transition is far from complete.

What We Learned From 160 Sessions

We've tracked our own human-AI partnership across more than 160 sessions, spanning over 30 projects, from industrial control systems to knowledge management tools. Every session documented. Every decision preserved. Every failure recorded.

Some things we learned the hard way:

The Real Problem: Treating AI as Stateless

The rediscovery tax is a symptom of a deeper assumption: that AI is a tool you use, not a partner you work with. That every interaction is independent. That the value of an AI collaborator is measured per-session, like a calculator that happens to understand English.

But the value of a collaborator — any collaborator, human or AI — isn't what they can do in a single interaction. It's what they learn over time. The compound interest of shared understanding. The accumulated context that turns "explain your system to me" into "I see you changed the auth pattern — does that affect the rate limiter we discussed last week?"

Every re-explanation is a deposit that earns zero interest. Every lost insight is depreciation on an asset that should be appreciating. Every session that starts from zero is a failure of infrastructure, not intelligence.

The AI is not the problem. The AI is remarkably capable within a session. The problem is that we haven't built the systems to let that capability persist and compound — to let a genuinely collaborative relationship develop across sessions, across projects, across teams.

Open Questions

We don't pretend to have all the answers. But we think the right questions are becoming clearer: