Top
Best
New

Posted by maheshrijal 4/14/2025

GPT-4.1 in the API(openai.com)
680 points | 492 commentspage 5
wongarsu 4/14/2025|
Is the version number a retcon of 4.5? On OpenAI's models page the names appear completely reasonable [1]: The o1 and o3 reasoning models, and non-reasoning there is 3.5, 4, 4o and 4.1 (let's pretend 4o makes sense). But that is only reasonable as long as we pretend 4.5 never happened, which the models page apparently does

1: https://platform.openai.com/docs/models

thund 4/15/2025||
Hey OpenAI if you ever need a Version Engineer, I’m available.
nsoonhui 4/14/2025||

  We will also begin deprecating GPT‑4.5 Preview in the API, as GPT‑4.1 offers improved or similar performance on many key capabilities at much lower cost and latency. GPT‑4.5 Preview will be turned off in three months
Here's something I just don't understand, how can ChatGPT 4.5 be worse than 4.1? Or the only thing bad is that the OpenAI naming ability?
chr15m 4/14/2025|
They tried something and it didn't work well. Branching paths of experimentation is not compatible with number-goes-up versioning.
neal_ 4/14/2025||
The better the benchmarks, the worse the model is. Subjectively for me the more advanced models dont follow instructions, and are less capable of implementing features or building stuff. I could not tell a difference in blind testing SOTA models gemini, claude, openai, deepseek. There has been no major improvements in the LLM space since the original models gained popularity. Each release claims to be much better the last, and every time i have been disappointed and think this is worse.

First it was the models stopped putting in effort and felt lazy, tell it to do something and it will tell you to do it your self. Now its the opposite and the models go ham changing everything they see, instead of changing one line, SOTA models rather rewrite the whole project and still not fix the issue.

Two years back I totally thought these models are amazing. I always would test out the newest models and would get hyped up about it. Every problem i had i thought if i just prompt it differently I can get it to solve this. Often times i have spent hours prompting starting new chats, adding more context. Now i realize its kinda useless and its better to just accept the models where they are, rather then try and make them a one stop shop, or try to stretch capabilities.

I think this release I won’t even test it out, im not interested anymore. I’ll probably just continue using deepseek free, and gemini free. I canceled my openai subscription like 6 months ago, and canceled claude after 3.7 disappointment.

composableaide 4/15/2025||
Excited to see 4.1 in the API. The Nano model pricing is comparable to Gemini Flash but not where we would like it to be: https://composableai.de/openai-veroeffentlicht-4-1-nano-als-...
forbiddenvoid 4/14/2025||
Lots of improvements here (hopefully), but still no image generation updates, which is what I'm most eager for right now.
taikahessu 4/14/2025||
Or text to speech generation ... but I guess that is coming.
dharmab 4/14/2025||
Yeah, I tried the 4o models and they severely mispronounced common words and read numbers incorrectly (eg reading 16000 as 1600)
Tinkeringz 4/14/2025||
They just realised a new image generation a couple of weeks ago, why are you eager for another one so soon?
nanook 4/14/2025||
Are the image generation improvements available via API? Don't think so
flakiness 4/14/2025||
Big focus on coding. It feels like a defensive move against Claude (and more recently, Gemini Pro) which became very popular in that regime. I guess they recently figured out some ways to train the model for these "agentic" coding through RL or something - and the finding is too new to apply 4.5 on time.
sc077y 4/15/2025||
I'm wondering if one of the big reasons that OpenAI is making gpt-4.5 deprecated is not only because it's not cost-effective to host but because they don't want their parent model being used to train competitors' models (like deepseek).
asdev 4/14/2025|
it's worse than 4.5 on nearly every benchmark. just an incremental improvement. AI is slowing down
usaar333 4/14/2025||
Or OpenAI is? After using Gemini 2.5, I did not feel "AI is slowing down". It's just this model isn't SOTA.
Nckpz 4/14/2025|||
They don't disclose parameter counts so it's hard to say exactly how far apart they are in terms of size, but based on the pricing it seems like a pretty wild comparison, with one being an attempt at an ultra-massive SOTA model and one being a model scaled down for efficiency and probably distilled from the big one. The way they're presented as version numbers is business nonsense which obscures a lot about what's going on.
conradkay 4/14/2025|||
It's like 30x cheaper though. Probably just distilled 4.5
GaggiX 4/14/2025|||
It's better on AIME '24, Multilingual MMLU, SWE-bench, Aider’s polyglot, MMMU, ComplexFuncBench while being much much cheaper and smaller.
asdev 4/14/2025||
and it's worse on just as many benchmarks by a significant amount. as a consumer I don't care about cheapness, I want the maximum accuracy and performance
GaggiX 4/14/2025||
As a consumer you care about speed tho, and GPT-4.5 is extremely slow, at this point just use a reasoning model if you want the best of the best.
HDThoreaun 4/14/2025|||
Maybe progress is slowing down but after using gemini 2.5 there clearly is still a lot being made.
simianwords 4/14/2025||
Sorry what is the source for this?
More comments...