r/LocalLLaMA · · 2 min read

Deepseek V4's 1M context window: the breaking point

Mirrored from r/LocalLLaMA for archival readability. Support the source by reading on the original site.

Just ran to verify deepseek v4's context claim of 1M and ran it across three production codebases like 45k (microservice), 180k (monorepo backend) and 520k(full stack app). For the observation, tasks included dependency tracing, cross file refractors and bug isolation to see where recall keeps up

under 150k

Got a solid performance like at 45k tokens, function calls traced across 8 files maintain accurate path reconstruction. At 180k, multi file refractors spanning 14 files show consistent architectural understand and no contradictions or context loss patterns

past 300k

precision quality degrades here. asked for exact line numbers from functions defined 400k tokens earlier, responses give "around line 230" instead of the actual 247. at 520k outputs shift to architectural summaries that skip implementation details, thats a problem if edge cases are a concern

the latency gap

Time to first token measures around 1.19s on deepinfra fp4 endpoint. Time to first answer in max reasoning mode stretches to around 120 seconds since the model completes internal chain of thought before producing visible output, which is really crticial for interative workflows to account for

provider benchmarks show 94% hallucination rate on unknown asnwer tasks (aa-omniscience) but v4 generates confident responses without even actual info. Shows up as references to nonexistent utility functions or phantom dependencies

on unknown answer tasks v4 generates confident responses without actual grounding, shows up as references to nonexistent utility functions or phantom dependencies. needs a validation layer for anything production critical

practical range

150-250k tokens appears optimal for coding work. full context retention, sub 2s response latency, minimal precision loss. past 300k requires defensive prompting and source verification.

the 1m window functions technically but needs careful handling tho. context size shifts which prompt engineering techniques matter rather than eliminating the need completely

submitted by /u/TangeloOk9486
[link] [comments]

Discussion (0)

Sign in to join the discussion. Free account, 30 seconds — email code or GitHub.

Sign in →

No comments yet. Sign in and be the first to say something.

More from r/LocalLLaMA