The 128K Token Lie
Most production LLM providers claim 128K token context windows. In practice, quality degrades sharply past 50K tokens, and nobody tells you why until you've already burned through your API budget.
I'm talking about real-world behavior: GPT-4 Turbo, Claude 3.5 Sonnet, Gemini 1.5 Pro all exhibit this. You stuff 100K tokens of documentation into the context, ask a question about something buried at token 60K, and the model either hallucinates, ignores it entirely, or gives you a vague non-answer. The advertised window is there — the model doesn't error out — but the useful window is roughly half.
This isn't just anecdotal frustration. The "lost in the middle" phenomenon, documented by Liu et al. (2023) in their NeurIPS paper, shows that transformer attention mechanisms struggle with information retrieval when the relevant context is buried deep in a long sequence. Models perform best on information at the start and end of the context, with a significant accuracy drop in the middle 40-60% of the window.
But the real production killer isn't just accuracy degradation. It's the interaction between context length, latency, and cost.
Continue reading the full article on TildAlice

Top comments (0)