Research Update — Improved Faithfulness Across the Full 200K Context Window
Anthropic's research team has published findings on improvements to Claude's faithfulness when operating across the full 200K token context window. A known limitation of large-context language models is the tendency for information retrieval quality to degrade for content that appears in the middle of the context — sometimes described informally as the "lost in the middle" problem. The research shows that targeted training improvements in Claude 4.5 Sonnet and Opus have substantially reduced this effect compared to the 3.x generation.
Key findings
- Middle-of-context recall — on a standardised multi-document retrieval evaluation, Sonnet 4.5 achieves 91% recall of information placed in the middle third of a 200K context, compared to 74% for Claude 3.7 Sonnet on the same evaluation
- Citation accuracy — when asked to cite specific passages from long documents, Claude 4.5 Sonnet produces accurate citations in 88% of cases versus 71% for the previous generation
- Instruction adherence at depth — complex multi-part instructions embedded deep within long contexts are followed at near the same rate as the same instructions at the start of the context
The improvements are most material for enterprise use cases involving long legal documents, technical specifications, and multi-party contract analysis where information is distributed throughout large documents rather than concentrated at the beginning.