logoalt Hacker News

OpenAI releases GPT-5.5 and GPT-5.5 Pro in the API

239 pointsby arabicaloriesyesterday at 6:23 PM134 commentsview on HN

GPT-5.5 - https://news.ycombinator.com/item?id=47879092 - April 2026 (1010 comments)


Comments

wincyyesterday at 7:43 PM

Just tried it out for a prod issue was experiencing. Claude never does this sort of thing, I had it write an update statement after doing some troubleshooting, and I said “okay let’s write this in a transaction with a rollback” and GPT-5.5 gave me the old “okay,

BEGIN TRAN;

-- put the query here

commit;

I feel like I haven’t had to prod a model to actually do what I told it to in awhile so that was a shock. I guess that it does use fewer tokens that way, just annoying when I’m paying for the “cutting edge” model to have it be lazy on me like that.

This is in Cursor the model popped up and so I tried it out from the model selector.

show 5 replies
guilamuyesterday at 8:05 PM

Just tested it on my homemade Wordpress+GravityForms benchmark and it's one of the worst model of the leaderboard performance wise and the worst value wise: https://github.com/guilamu/llms-wordpress-plugin-benchmark

I know it's only on a single benchmark, but I dont understand how it can be so bad...

show 5 replies
Topfiyesterday at 9:18 PM

Pricing by context length:

Input: $5/M tokens at <=272K, $10/M tokens above 272K.

Output: $30/M tokens at <=272K, $45/M tokens above 272K.

Cache read: $0.50/M tokens at <=272K, $1/M tokens above 272K.

Significantly more expensive than Opus 4.7 beyond 272K and at least in my tasks, I haven't seen the model that much more token efficient, certainly not to such a degree that it'd compensate this difference. GPT-5.4 had a solid context window at 400k with reliable compaction, both appear somewhat regressed, though still to early to truly say whether compaction is less reliable. Also, I have found frontend output to still skew towards that one very distinct, easily noticeable, card laden, bluesy hue overindulged template that made me skeptical of Horizon Alpha/Beta pre GPT-5s release. Ended up doing amazing at the time for task adherence, which made it very useful for me outside that one major deficit. The fact that GPT-5.5 is still so restricted in that area is weird considering it's supposed to be an entirely new foundation.

jvidalvyesterday at 10:21 PM

In the only one that feels that OpenAI has bots/commenters on payroll on all this kind of news downplaying Claude and stating how much better Codex is?

There is too much and there are too many, and some of their takes don’t fly if you use Claude daily.

show 3 replies
sigmoid10yesterday at 6:56 PM

Huh. Yesterday they said:

>API deployments require different safeguards and we are working closely with partners and customers on the safety and security requirements for serving it at scale.

And now this. I guess one day counts as "very soon." But I wonder what that meant for these safeguards and security requirements.

show 3 replies
nltoday at 12:08 AM

Second model to get 25/25 on my benchmark (after Opus 4.7): https://sql-benchmark.nicklothian.com/?highlight=openai_gpt-...

Cheaper and slower than Opus.

neosatyesterday at 7:19 PM

Enterprise user here and still seeing only 5.4. Yesterday's announcement said that it will take a few hours to roll out to everybody. OpenAI needs better GTM to set the right expectations.

show 2 replies
gertlabsyesterday at 11:45 PM

Comprehensive coding reasoning benchmark results for GPT 5.5 with max reasoning are up at https://gertlabs.com/

Live decision and heavier agentic evals will continue being uploaded for 24 hours but I don't expect its leaderboard position to change at this point.

GPT 5.5 is the most intelligent public model. And significantly faster than its predecessor.

robertwt7yesterday at 9:37 PM

Gpt 5.5 combined with codex is really good. I actually have no doubt whenever I asked questions, plan, or implement a code with it. With opus 4.7, I have to keep double checking because it doesnt follow the CLAUDE.md instruction, it hallucinates a lot, by default it makes things up when it can’t find the answer to something. Its crazy how quickly people are saying that OpenAI is left behind last year when they declared code red and look at where we are now

czkyesterday at 7:16 PM

API page lists the knowledge cutoff as Dec 01, 2025 but when prompting the model it says June 2024.

   Knowledge cutoff: 2024-06
   Current date: 2026-04-24

   You are an AI assistant accessed via an API.
show 5 replies
zerof1lyesterday at 9:12 PM

I don't see any meaningful performance improvements in those paid models anymore.

They all roughly produce junior developer-level code, continue to have mental breakdowns in their “thinking” stage, occasionally hallucinate things, delete pieces of code/docs they don’t understand or don’t like, use 1.5 times the necessary words to explain things when generating docs and so on.

I'm now testing "avoid sycophancy, keep details short and focus on the facts" in my AGENTS.md files.

show 2 replies
ftononyesterday at 8:06 PM

Looks like the default config in the chat is instant 5.3, it only uses the 5.5 on the thinking variant

show 1 reply
QuadrupleAyesterday at 8:39 PM

Exactly double the cost of GPT 5.4 - $5 per MTok input, $0.50 cached, $30 output.

All the AI players definitely seem to be trying to claw more money out of their users at the moment.

show 2 replies
redsaberyesterday at 7:07 PM

not available for Github Copilot pro(only in pro+, business and enterprise), I am really now feeling the era of subsidized AI is over.

show 2 replies
throw03172019yesterday at 6:48 PM

Faster than anticipated because of Deepseek release?

show 4 replies
jedisct1today at 8:26 AM

    $ uvx swival --provider chatgpt --model gpt-5.5

    APIError: ChatgptException
Ok, still not available everywhere apparently :(
pants2yesterday at 6:55 PM

Is anyone here actually using pro models through the API? I'd be very curious what the use-case is.

show 2 replies
_pdp_yesterday at 8:12 PM

A very expensive model for API usage. Fine in codex I think.

gigatexalyesterday at 7:24 PM

what's the real world comparison to opus 4.7 fellow coders?

show 1 reply
pillefitzyesterday at 7:43 PM

Please consider the ethical aspects of giving money to OpenAI versus alternatives.

show 1 reply
refulgentisyesterday at 9:14 PM

I'm absolutely stunned by what I've seen from 5.5. I thought it'd be a nothingburger and ~= Opus.

Gave it two very long-running problems I haven't had the courage to work on in the last 2.5 years, solved each within an hour.

- An incremental streaming JSON decoder that can optionally take a list of keys to stop decoding after. 1800 LOC about 30 minutes later, and now my local-first apps first sync time is 0.8s instead of 75s when there's 1.5 GB of data locally.

- Flutter Web can compile to WASM and then render via Skia WASM. I've been getting odd crashes during rapid animation for months. In an hour, it got Skia WASM checked out, building locally, a Flutter test script, and root caused the issue to text shadows and font glyphs (technically, not solved yet, I want to get to the point we have Skia / Flutter patch(es))

If you told me a week ago that an LLM could do either of these, without heavy guidance, I'd be stunned. And I regularly push them to limits, ex. one of Opus' last projects was a tolerant JSON decoder, and it ended up being 8% faster than the one built-in to Dart/Flutter, which has plenty of love and attention. (we're cheating a little, that's why it's faster. TL;DR: LLMs will emit control characters in JSON and that's fine for me, treating them as fine means file edit error rates go from ~2% to 0%)

I just wish it was cheaper, but, don't we all...

Jhonwilsonyesterday at 7:41 PM

that is great news

willjyesterday at 9:39 PM

[dead]

benjiro3000yesterday at 8:14 PM

[dead]

woohinyesterday at 9:48 PM

[dead]

theo_park87yesterday at 7:36 PM

[dead]

XCSmeyesterday at 9:01 PM

GPT 5.5 is close to Opus 4.7, but at 7x the cost[0]...

Either Opus 4.7 miscounts reasoning tokens, or it's A LOT more efficient than GPT 5.5

I thought they made GPT 5.5 more token efficient than 5.4, but it uses 2x the reasoning tokens.

[0]: https://aibenchy.com/compare/openai-gpt-5-5-medium/openai-gp...

rvnxyesterday at 7:11 PM

Very bad habit these safeguards. These "safety" filters are counter-productive and even can be dangerous.

In my place for example, a lot of doctors are using ChatGPT both to search diagnosis and communicate with non-English speaking patients.

Even yourself, when you want to learn about one disease, about some real-world threats, some statistics, self-defense techniques, etc.

Otherwise it's like blocking Wikipedia for the reason that using that knowledge you can do harmful stuff or read things that may change your mind.

Freedom to read about things is good.

show 2 replies