Top
Best
New

Posted by mudkipdev 8 hours ago

GPT-5.4(openai.com)
https://openai.com/index/gpt-5-4-thinking-system-card/

https://x.com/OpenAI/status/2029620619743219811

594 points | 524 commentspage 4
motbus3 6 hours ago|
Sam Altman can keep his model intentionally to himself. Not doing business with mass murderers
dandiep 7 hours ago||
Anyone know why OpenAI hasn't released a new model for fine tuning since 4.1? It'll be a year next month since their last model update for fine tuning.
zzleeper 7 hours ago||
For me the issue is why there's not a new mini since 5-mini in August.

I have now switched web-related and data-related queries to Gemini, coding to Claude, and will probably try QWEN for less critical data queries. So where does OpenAI fits now?

Rapzid 5 hours ago|||
Also interested in this and a replacement for 4.1/4.1-mini that focuses on low latency and high accuracy for voice applications(not the all-in-one models).
qoez 7 hours ago||
I think they just did that because of the energy around it for open source models. Their heart probably wasn't in it and the amount of people fine tuning given the prices were probably too low to continue putting in attention there.
petetnt 1 hour ago||
Whoa, I think GPT-5.3 Instant was a disappointment, but GPT-5.4 is definitely the future!
jcmontx 7 hours ago||
5.4 vs 5.3-Codex? Which one is better for coding?
embedding-shape 7 hours ago||
Literally just released, I don't think anyone knows yet. Don't listen to people's confident takes until after a week or two when people actually been able to try it, otherwise you'll just get sucked up in bears/bulls misdirected "I'm first with an opinion".
vtail 7 hours ago|||
Looking at the benchmarks, 5.4 is slightly better. But it also offers "Fast" mode (at 2x usage), which - if it works and doesn't completely depletes my Pro plan - is a no brainer at the same or even slightly worse quality for more interactive development.
Someone1234 7 hours ago|||
Related question:

- Do they have the same context usage/cost particularly in a plan?

They've kept 5.3-Codex along with 5.4, but is that just for user-preference reasons, or is there a trade-off to using the older one? I'm aware that API cost is better, but that isn't 1:1 with plan usage "cost."

awestroke 7 hours ago|||
Opus 4.6
jcmontx 6 hours ago|||
Codex surpassed Claude in usefulness _for me_ since last month
baal80spam 5 hours ago|||
[flagged]
esafak 7 hours ago||
For the price, it seems the latter. I'd use 5.4 to plan.
jeff_antseed 6 hours ago||
The 1M context vs compaction tradeoff is interesting from a routing angle too — longer context requests are fundamentally more expensive per request, which changes which provider wins on a P2P inference market.

A model like this shifts routing decisions: for tasks where 1M context actually helps (reverse engineering, large codebase analysis), you'd want to route to a provider who's priced for that workload. For most tasks, shorter context + cheaper model wins.

The routing layer becomes less about "pick the best model" and more about "pick the best model for this specific task's cost/quality tradeoff." That's actually where decentralized inference networks (building one at antseed.com) get interesting — the market prices this naturally.

freedomben 2 hours ago||
> When toggled on, /fast mode in Codex delivers up to 1.5x faster token velocity with GPT‑5.4. It’s the same model and the same intelligence, just faster.

I hate these blog posts sometimes. Surely there's got to be some tradeoff. Or have we finally arrived at the world's first "free lunch"? Otherwise why not make /fast always active with no mention and no way to turn it off?

paxys 7 hours ago||
"Here's a brand new state-of-the-art model. It costs 10x more than the previous one because it's just so good. But don't worry, if you don't want all this power you can continue to use the older one."

A couple months later:

"We are deprecating the older model."

OutOfHere 7 hours ago|
That's a misrepresentation of the cost. It is simply false. The cost is noted here: https://news.ycombinator.com/item?id=47265144
atkrad 3 hours ago||
What is the main difference between this version with the previous one?
butILoveLife 5 hours ago||
Anyone else completely not interested? Since GPT5, its been cost cutting measure after cost cutting measure.

I imagine they added a feature or two, and the router will continue to give people 70B parameter-like responses when they dont ask for math or coding questions.

alpineman 7 hours ago|
No thanks. Already cancelled my sub.
More comments...