logoalt Hacker News

conceptionyesterday at 10:16 PM4 repliesview on HN

I noticed 1M context window is default and no way not to use it. If your context is at 500-900k tokens every prompt, you’re gonna hit limits fast.


Replies

Wowfunhappyyesterday at 10:45 PM

I had to double check that they'd removed the non-1M option, and... WTF? This is what's in `config` → `model`

    1. Default (recommended)   Opus 4.6 with 1M context · Most capable for complex work
    2. Sonnet                   Sonnet 4.6 · Best for everyday tasks
    3. Sonnet (1M context)      Sonnet 4.6 with 1M context · Billed as extra usage · $3/$15 per Mtok
    4. Haiku                    Haiku 4.5 · Fastest for quick answers
So there's an option to use non-1M Sonnet, but not non-1M Opus?

Except wait, I guess that actually makes sense, because it says Sonnet 1M is billed as extra usage... but also WTF, why is Sonnet 1M billed as extra usage? So Opus 1M is included in Max, but if you want the worse model with that much context, you have to pay extra? Why the heck would anyone do that?

The screen does also say "For other/previous model names, specify with --model", so maybe you can use that to get 200K Opus, but I'm very confused why Anthropic wouldn't include that in the list of options. What a strange UX decision.

aberohamyesterday at 10:19 PM

export CLAUDE_CODE_DISABLE_1M_CONTEXT=1

show 1 reply
nextaccountictoday at 1:35 AM

do you pay for the full context every prompt? what happened with the idea of caching the context server side?

show 2 replies
zhangchentoday at 1:37 AM

[dead]