Sourcegraph’s Amp Tries a New Fix for the Long-Conversation Problem

Written by ainativedev | Published 2026/01/03
Tech Story Tags: ai | sourcegraph | sourcegraph-amp | ai-long-context-drift | long-context-models | long-context-drift | sourcegraph-handoff | ai-native-development

TLDRAmp's new "handoff" feature replaces compaction by packaging relevant context into new threads while navigating complex discussions.via the TL;DR App

When conversations with AI stretch too far, the model starts to lose track — earlier messages fade from its memory as the conversation grows, and meaning can become garbled. While larger context windows go some way toward solving this problem, they merely delay that steady drift in understanding.

Amp, the coding assistant from Sourcegraph, is trying to fix that by changing how its users move from one idea to the next. The company has replaced its long-standing “compaction” feature with something new: Handoff — a tool meant to carry context forward without dragging along the entire past.

Until now, Amp gave users a way to manage long threads manually, by running a compaction (/compact) command that condensed the history into a short summary before continuing from there. It was a practical fix for a familiar problem: as conversations stretched on, the assistant would lose focus and become muddled. Compacting a thread trimmed that excess, freeing up space in the model’s limited memory window and restoring some sharpness to its replies. In effect, it let users perform a quick cleanup without abandoning the thread altogether — a reset of attention rather than intent.

Handoff, according to Amp, takes a different approach. Rather than condensing the past, it creates a new thread from the old one — packaging the relevant context, setting a clear goal, and letting users review what gets passed on.

The result is a sequence of linked steps, each with its own focus.

How handoff works

When invoked — typically by running the /handoff command in the Amp CLI or selecting Handoff from the thread menu in the interface — the feature lets you specify a goal for the new thread. Amp then analyzes the current discussion to generate a draft prompt, along with a list of files or artifacts it thinks are relevant. In the editor extensions, you can also access Handoff directly from the token usage hover, making the transition available without leaving your coding environment.

  • /handoff now implement this for teams as well, not just individual users
  • /handoff execute phase one of the created plan
  • /handoff check the rest of the codebase and find other places that need this fix

In the example below, a user finishes mapping out an implementation plan in Amp, then uses the /handoff command to move that context into a fresh thread. The assistant analyzes the existing discussion, generates a draft prompt for the next phase, and opens a new workspace for implementation — leaving the original plan untouched.

The Codex compaction concern

A recent internal report from OpenAI offers a useful data point for understanding why Amp moved away from compaction.

Among several findings, the Codex team noted that its automated compaction system — which summarized a session and restarted it whenever the model’s context window neared its limit — was contributing to a gradual decline in performance over time, a trend that emerged soon after the arrival of GPT-5 Codex in September 2025. As sessions accumulated more “compacts,” accuracy fell, and “recursive summaries” (i.e. summaries of summaries) began to distort earlier reasoning.

OpenAI’s engineers traced part of the problem to the very mechanism meant to keep conversations continuous. The report included a chart showing how the share of Codex sessions involving compaction more than doubled through October.

Indeed, in the wake of OpenAI’s report publication, “Dan Mac,” a senior engineer at Amp, drew the parallel bluntly.

“You should basically never use compaction,” he wrote on social media. “Thank me later.”

One commenter framed the decision as a question of taste. Amp, they noted, is giving users control over what context moves forward through Handoff, but (notably) not which model handles it — a deliberate constraint that favors consistency and simplicity over full flexibility. It’s a trade-off between power and polish, and, as the commenter put it, “now the market decides."

The context correction

Ultimately, Amp’s move sits within a broader rethinking of how AI systems handle context. Across the industry, tools are being rebuilt to measure, extend, or externalize memory, such as the new open source Context-Bench benchmark for “agentic context engineering,” Anthropic’s agent skills layer for Claude Code, and ecosystem experiments in generating contextual scaffolding around tasks. Each tackles the same problem from a different angle: how to preserve continuity without distortion.

Amp’s Handoff joins that trend, not by stretching context windows further, but by teaching systems — and users — how to let go of them cleanly.



Written by ainativedev | Your source for the latest in AI Native Development — news, insights, and real-world developer experiences.
Published by HackerNoon on 2026/01/03