Top
Best
New

Posted by hansonw 11/19/2025

Building more with GPT-5.1-Codex-Max(openai.com)
483 points | 319 commentspage 4
syntaxing 11/19/2025|
I rarely used Codex compared to Claude because it was extremely slow in GitHub copilot . Like maybe 2-5X slower than Claude Sonnet. I really wish they just made their models faster than “better”
levocardia 11/19/2025||
Very interesting to see the range of peoples' preferences. I would almost always prefer smart over fast; I have all my LLMs to be all-thinking-all-the-time.
mrguyorama 11/19/2025|||
GPT-5 was recently updated to make it more "thinking" and "warmer" or whatever and now a task (semantically compare these two short files) that used to take 5 seconds and reliably produce useful and consistent output now takes 90 seconds to "think" (while it's thinking output makes it pretty clear there is zero thinking happening) and produces a completely differently structured output every single time, making the tool not only slower and more expensive to use, but worse at a simple task that LLMs should be very good at.

There's an option to "get a quick answer" and I hoped clicking that would revert to previous performance and instead what it does is ignore that I uploaded two files and asks me to upload the files

Literally the only real good task I've found for these dumb things and they still found a way to fuck it up because they need to keep the weirdos and whales addicted. It's now almost easier to go back to comparing these files by eye, or just bite the bullet and finally write a few lines of python to actually do it right and reliably.

syntaxing 11/19/2025||||
It’s a balance, I haven’t felt like codex provided anything that Sonnet 4.5 didn’t. Why wait longer for getting the same results.

Though that does bring up an interesting point. Anecdotally, Sonnet does a lot more grep-ing while Codex reads files straight up. Might be the difference in speed and maybe smarter models will do better. Once this model is on copilot, I can test it out.

theshrike79 11/21/2025|||
Depends on what I'm doing.

If I'm adding a new feature I want to test and keep the flow going, I want fast -> Claude.

But if I want a report on test coverage or possible security issues or a bigger refactoring, I want slow careful and smart -> Codex.

jasonsb 11/19/2025|||
OpenAI doesn't want you to use their models outside of their own products, which is why the API and integrations like Github Copilot are super slow.
sumedh 11/19/2025||
That does not make business sense though. If people want to use Open AI models in Copilot and other tools and they dont perform they will just switch to another model and not come back they are not going to use Codex.
nartho 11/19/2025||
Have you tried Mistral ? Definitely one of the fastest models
theshrike79 11/21/2025|||
I've tried Mistral for coding and it seems to be laughably bad every time. Dunno what I'm doing wrong.
syntaxing 11/19/2025|||
My employer doesn’t offer/allow anything besides the “traditional” offerings on GitHub copilot.
esafak 11/19/2025||
How efficient is it; does it go through your subscription quota faster?
kytazo 11/19/2025||
500 Internal Server Error.
morog 11/19/2025|
ditto. Also OpenAI vector stores are down right now across the board
kachapopopow 11/19/2025||
not sure if I am actually using 5.1-codex-max or just normal 5.1-codex (is there even 5.1-codex?) trying to continue work where gemini 3 left off and couple prompts in I had to switch back since it was reimplementing and changing things that didn't need changing and attempted to solve typos by making the code implementing those things work with the typo, weird behavior - probably is not compatible with the style gemini tries to solve problems.
sumedh 11/19/2025|
Just run the /model command in codex and select the model which you want.
LZ_Khan 11/19/2025||
Woah, metr results look impressive. Still looking exponential
andai 11/19/2025||
The graph showing higher performance for fewer thinking tokens is really interesting!

It would be even more interesting to see how Sonnet and Haiku compare with that curve.

bgwalter 11/19/2025||
So they all release before the Nvidia numbers tonight. The real question is: How well can Nvidia hide the circular deals in the books?
nowittyusername 11/19/2025||
Glad to see evolution of proper context management. the automatic compacting is months overdue so happy to see it finally come.
ed_mercer 11/20/2025|
As a long time CC user, I was like "Wait, they didn't have auto-compaction all this time??"
LordIsBack 11/23/2025||
I don't think GPT-5.1-Codex-Max is better than Sonnet 4.5 still.
More comments...