logoalt Hacker News

Topfiyesterday at 9:18 PM0 repliesview on HN

Pricing by context length:

Input: $5/M tokens at <=272K, $10/M tokens above 272K.

Output: $30/M tokens at <=272K, $45/M tokens above 272K.

Cache read: $0.50/M tokens at <=272K, $1/M tokens above 272K.

Significantly more expensive than Opus 4.7 beyond 272K and at least in my tasks, I haven't seen the model that much more token efficient, certainly not to such a degree that it'd compensate this difference. GPT-5.4 had a solid context window at 400k with reliable compaction, both appear somewhat regressed, though still to early to truly say whether compaction is less reliable. Also, I have found frontend output to still skew towards that one very distinct, easily noticeable, card laden, bluesy hue overindulged template that made me skeptical of Horizon Alpha/Beta pre GPT-5s release. Ended up doing amazing at the time for task adherence, which made it very useful for me outside that one major deficit. The fact that GPT-5.5 is still so restricted in that area is weird considering it's supposed to be an entirely new foundation.