Context Rot: How increasing input tokens impacts LLM performance

Published:Jul 14, 2025 19:25
1 min read
Hacker News

Analysis

The article discusses the phenomenon of 'context rot' in LLMs, where performance degrades as the input context length increases. It highlights that even state-of-the-art models like GPT-4.1, Claude 4, Gemini 2.5, and Qwen3 are affected. The research emphasizes the importance of context engineering, suggesting that how information is presented within the context is crucial. The article provides an open-source codebase for replicating the results.

Reference

Model performance is non-uniform across context lengths, including state-of-the-art GPT-4.1, Claude 4, Gemini 2.5, and Qwen3 models.