|
In my first two posts of our AI series, I explored how today’s AI coding assistants bump into memory gaps when context windows are too short, and why code-generation experiments often feel like working with a very smart, but distractible, junior teammate. In this third installment, I look at the flip side: the problems that arise when context windows are too long. The culprit is a phenomenon researchers call “context rot”. By, Bill Miller What Is “Context Rot”?
Researchers at Chroma recently described the problem in detail under the label ‘context rot’ (https://research.trychroma.com/context-rot). The term refers to what happens when a long conversation with an AI drifts off course because the model clings to outdated or irrelevant parts of the discussion. If you’ve ever been deep in a project—planning a product launch, iterating on a financial model, or debugging code—you know that directions change. Dead ends appear. Teams pivot. But AI systems don’t always adapt gracefully. Even after you explicitly abandon an approach, the model may later “resurrect” those discarded ideas and weave them back into its answers. These abandoned threads become “distractions” - leftovers from old paths that dilute the clarity of the current work. Why It Happens AI models like ChatGPT and Claude are trained to continue conversations by drawing from the “entire context window” of prior messages. They don’t “know” which parts are still relevant—they only know that those words are in scope. The result:
A Simple Example In my own use, I’ve seen this play out when working with ChatGPT-5 (thinking mode) and the latest versions of Claude Code. Imagine starting a coding project where you explore “Approach A,” realize it won’t work, and pivot to “Approach B.” Even after making that switch explicit, the model sometimes falls back on snippets or assumptions from Approach A. It’s not malicious or careless - it’s simply doing what it’s designed to do: predict the next best word given “everything” it has seen. But for you, the human trying to move forward, it feels like your collaborator has forgotten the decision to pivot. When I instead start a “fresh conversation” with just the relevant context for Approach B, the results are far cleaner and more aligned. Managing Context Rot in Practice Until the models get better at distinguishing “current” from “abandoned” context, there are a few strategies you can use: 1. Start fresh when directions change. If you’ve pivoted significantly, spin up a new chat and provide only the context you want carried forward. 2. Summarize the new direction. Give the AI a clear, short recap of where you are now, almost like a project reset. 3. Use external memory tools. Some emerging platforms allow you to pin or filter context (think of them as “working documents” the AI can refer back to). 4. Treat AI like a collaborator who takes notes too literally. Don’t assume it “knows what you mean”—spell out when something should be ignored going forward. Why This Matters for Executives For executives experimenting with AI in workflows—strategy, sales, finance, product—the lesson is this:
Looking Ahead: Private Orchestration and Memory Control This is where private orchestration engines like Kamiwaza, a FirstMile Ventures portfolio company, point the way forward. Beyond the public inferencing options everyone has access to, Kamiwaza gives teams much finer-grained control over how inference is managed, including direct control over the model’s “working scratchpad” or what AI engineers call the “KV cache”.. Think of this scratchpad as the AI’s short-term notebook of the conversation:
Taken together, scratchpad control, disciplined prompting, and model selection enable conversations that remain coherent, efficient, and free from distraction. For executives, the message is clear: infrastructure choices—not just which AI model you pick—will determine whether your AI collaborator stays sharp or drifts into context rot. In short: AI doesn’t yet know what to forget. But with private orchestration and smarter memory management, we’re starting to build systems that can. Comments are closed.
|