logoalt Hacker News

damstayesterday at 8:12 PM4 repliesview on HN

There is extra cost for >272K:

> For models with a 1.05M context window (GPT-5.4 and GPT-5.4 pro), prompts with >272K input tokens are priced at 2x input and 1.5x output for the full session for standard, batch, and flex.

Taken from https://developers.openai.com/api/docs/models/gpt-5.4


Replies

WXLCKNOtoday at 12:24 AM

Anthropic literally don't allow you to use the 1M context anymore on Sonnet and Opus 4.6 without it being billed as extra usage immediately.

I had 4.5 1M before that so they definitely made it worse.

OpenAI at least gives you the option of using your plan for it. Even if it uses it up more quickly.

show 1 reply
minimaxiryesterday at 8:29 PM

Good find, and that's too small a print for comfort.

glensteinyesterday at 8:37 PM

Wow, that's diametrically the opposite point: the cost is *extra*, not free.

show 1 reply
fragmedeyesterday at 8:16 PM

Which, Claude has the same deal. You can get a 1M context window, but it's gonna cost ya. If you run /model in claude code, you get:

    Switch between Claude models. Applies to this session and future Claude Code sessions. For other/previous model names, specify with --model.
    
       1. Default (recommended)   Opus 4.6 · Most capable for complex work
       2. Opus (1M context)        Opus 4.6 with 1M context · Billed as extra usage · $10/$37.50 per Mtok
       3. Sonnet                   Sonnet 4.6 · Best for everyday tasks
       4. Sonnet (1M context)      Sonnet 4.6 with 1M context · Billed as extra usage · $6/$22.50 per Mtok
       5. Haiku                    Haiku 4.5 · Fastest for quick answers