Top
Best
New

Posted by meetpateltech 8 hours ago

GPT‑5.3‑Codex‑Spark(openai.com)
509 points | 206 commentspage 3
wxw 7 hours ago||
Great stuff. People are getting used to agents as the interface for everything, even work as simple as "change label X to label Y". More speed on that front is welcome. The Codex "blended mode" they refer to will be useful (similar to Claude Code bouncing between haiku and opus).

I imagine it's a win-win. This could significantly help their tokenomics.

The example showing a plan being generated instantaneously is interesting. Human understanding will end up as the last, true bottleneck.

mynti 6 hours ago||
With the rough numbers from the blog post at ~1k tokens a second in Cerebras it should put it right at the same size as GLM 4.7, which also is available at 1k tokens a second. And they say that it is a smaller model than the normal Codex model
Havoc 2 hours ago|
You can’t extrapolate size of model from speed that way. Architecture difference, load etc will screw up the approximation
dalemhurley 5 hours ago||
This is a win for agents, speed and intelligence is crucial to the loop. If the time and token cost is small you can iterate many times to correct mistakes.

Got to wonder why Wall Street is dumping NVIDIA.

SamDc73 5 hours ago|
I mean they are only running a small version of codex can they run the full one? Or the technology isn't there yet?
jbellis 3 hours ago||
really too bad that the codex models are so tightly coupled to the codex harness as to be useless for everything else

edit: not useless in a absolute sense, but worse than the vanilla gpt models

thehamkercat 3 hours ago|
GPT-5.2-codex or 5.3-codex Works pretty well for me in opencode
cjbarber 7 hours ago||
It'll be nice when there's smarter routing between models, or easier routing, so some things get sent to the fast model, some get sent to the cheap model, some get sent to the smart model, etc.
alexhans 7 hours ago||
When I saw Spark my mind went to Apache Spark and wondered if we were learning all the lessons in orchestration of driver/worker and data shuffling from that space.
hchak 6 hours ago|
Cerebras out here catching dubs. Does anyone know if Groq is running DGX Cloud inference or am I tripping?
More comments...