Watch AI Native DevCon on demandWatch AI Native DevCon on YouTube
Logo
  • Articles146
  • Podcast89
  • Devtools Landscape624
  • Events26
  • Newsletter36
  • DevCon
  • Articles146
  • Podcast89
  • Devtools Landscape624
  • Events26
  • Newsletter36
  • DevCon

Get Weekly Insights

Stay up to date with the latest in AI Native Development: insights, real-world experiences, and news from developers and industry leaders.

Email Address*
Full Name
Company
Company Role
We value your privacy. Your email will only be used for updates about AI Native Dev and Tessl.
Logo
  • Discord
  • LinkedIn
  • X
  • YouTube
  • Spotify
  • Apple Podcasts
  • Home
  • Articles
  • Podcast
  • Landscape
  • About
  • Privacy Policy
  • Code of Respect
  • Cookies
  • Contact
© AI Native Dev
Back to articlesAmp ditches compaction for ‘handoff’ to fix AI’s long-context drift

14 Nov 20256 minute read

Paul Sawers

Freelance tech writer at Tessl, former TechCrunch senior writer covering startups and open source

LinkedIn
X
Substack
AI Coding Tools
AI Tools & Assistants
Developer Experience
Agentic Systems
Technical Deep Dive
Table of Contents
How handoff works
Back to articles

Amp ditches compaction for ‘handoff’ to fix AI’s long-context drift

14 Nov 20256 minute read

When conversations with AI stretch too far, the model starts to lose track — earlier messages fade from its memory as the conversation grows, and meaning can become garbled. While larger context windows go some way toward solving this problem, they merely delay that steady drift in understanding.

Amp, the coding assistant from Sourcegraph, is trying to fix that by changing how its users move from one idea to the next. The company has replaced its long-standing “compaction” feature with something new: Handoff — a tool meant to carry context forward without dragging along the entire past.

Until now, Amp gave users a way to manage long threads manually, by running a compaction (/compact) command that condensed the history into a short summary before continuing from there. It was a practical fix for a familiar problem: as conversations stretched on, the assistant would lose focus and become muddled. Compacting a thread trimmed that excess, freeing up space in the model’s limited memory window and restoring some sharpness to its replies. In effect, it let users perform a quick cleanup without abandoning the thread altogether — a reset of attention rather than intent.

Handoff, according to Amp, takes a different approach. Rather than condensing the past, it creates a new thread from the old one — packaging the relevant context, setting a clear goal, and letting users review what gets passed on.

The result is a sequence of linked steps, each with its own focus.

How handoff works

When invoked — typically by running the /handoff command in the Amp CLI or selecting Handoff from the thread menu in the interface — the feature lets you specify a goal for the new thread. Amp then analyzes the current discussion to generate a draft prompt, along with a list of files or artifacts it thinks are relevant. In the editor extensions, you can also access Handoff directly from the token usage hover, making the transition available without leaving your coding environment.

Handoff in VS Code

In the Amp terminal, it might look something like this, with the user setting the direction for the next phase:

  • /handoff now implement this for teams as well, not just individual users
  • /handoff execute phase one of the created plan
  • /handoff check the rest of the codebase and find other places that need this fix

In the example below, a user finishes mapping out an implementation plan in Amp, then uses the /handoff command to move that context into a fresh thread. The assistant analyzes the existing discussion, generates a draft prompt for the next phase, and opens a new workspace for implementation — leaving the original plan untouched.

Handoff in the Amp CLI

The Codex compaction concern

A recent internal report from OpenAI offers a useful data point for understanding why Amp moved away from compaction.

Among several findings, the Codex team noted that its automated compaction system — which summarized a session and restarted it whenever the model’s context window neared its limit — was contributing to a gradual decline in performance over time, a trend that emerged soon after the arrival of GPT-5 Codex in September 2025. As sessions accumulated more “compacts,” accuracy fell, and “recursive summaries” (i.e. summaries of summaries) began to distort earlier reasoning.

OpenAI’s engineers traced part of the problem to the very mechanism meant to keep conversations continuous. The report included a chart showing how the share of Codex sessions involving compaction more than doubled through October.

Compaction frequency with Codex (Credit: OpenAI)

Amp’s new Handoff feature can be read as a quiet response to the same pattern, replacing the cycle of compression and re-summarization with a cleaner break between phases of work, carrying forward only what still matters.

Indeed, in the wake of OpenAI’s report publication, “Dan Mac,” a senior engineer at Amp, drew the parallel bluntly.

“You should basically never use compaction,” he wrote on social media. “Thank me later.”

One commenter framed the decision as a question of taste. Amp, they noted, is giving users control over what context moves forward through Handoff, but (notably) not which model handles it — a deliberate constraint that favors consistency and simplicity over full flexibility. It’s a trade-off between power and polish, and, as the commenter put it, “now the market decides."

The community reacts

The context correction

Ultimately, Amp’s move sits within a broader rethinking of how AI systems handle context. Across the industry, tools are being rebuilt to measure, extend, or externalize memory, such as the new open source Context-Bench benchmark for “agentic context engineering,” Anthropic’s agent skills layer for Claude Code, and ecosystem experiments in generating contextual scaffolding around tasks. Each tackles the same problem from a different angle: how to preserve continuity without distortion.

Amp’s Handoff joins that trend, not by stretching context windows further, but by teaching systems — and users — how to let go of them cleanly.

Resources

Visit resource
Amp
Visit resource
Amp launches Handoff
Visit resource
The new Context-Bench benchmark
Visit resource
Anthropic gives Claude Code contextual intelligence
Visit resource
Conare conjures context for Claude Code

Related Articles

Code, meet crowd: Amp bets on ‘social coding’ with new public profiles

10 Nov 2025

Amp’s new business model? Ad-supported AI coding

28 Oct 2025

OpenAI adds Safeguard to GPT-OSS, letting developers set their own safety rules

5 Nov 2025

Paul Sawers

Freelance tech writer at Tessl, former TechCrunch senior writer covering startups and open source

LinkedIn
X
Substack
AI Coding Tools
AI Tools & Assistants
Developer Experience
Agentic Systems
Technical Deep Dive
Table of Contents
How handoff works

Resources

Visit resource
Amp
Visit resource
Amp launches Handoff
Visit resource
The new Context-Bench benchmark
Visit resource
Anthropic gives Claude Code contextual intelligence
Visit resource
Conare conjures context for Claude Code

Related Articles

Code, meet crowd: Amp bets on ‘social coding’ with new public profiles

10 Nov 2025

Amp’s new business model? Ad-supported AI coding

28 Oct 2025

OpenAI adds Safeguard to GPT-OSS, letting developers set their own safety rules

5 Nov 2025