Top
Best
New

Posted by meetpateltech 12/17/2025

Gemini 3 Flash: Frontier intelligence built for speed(blog.google)
Docs: https://ai.google.dev/gemini-api/docs/gemini-3

Developer Blog: https://blog.google/technology/developers/build-with-gemini-...

Model Card [pdf]: https://deepmind.google/models/model-cards/gemini-3-flash/

Gemini 3 Flash in Search AI mode: https://blog.google/products/search/google-ai-mode-update-ge...

Deepmind Page: https://deepmind.google/models/gemini/flash/

1102 points | 580 commentspage 8
XCSme 12/18/2025|
LLMs are weird, Gemini 3 flash beats Gemini 3 Pro on some benchmarks (MMMU-PRO)
Def_Os 12/17/2025||
Consolidating their lead. I'm getting really excited about the next Gemma release.
lazto 12/18/2025||
Wow, this is really an amazing model, and the experience is truly stunning.
agentifysh 12/17/2025||
so hat's why logan posed 3 lightning emojis. at $0.50/M for input and $3.00/M for output, this will put serious pressure on OpenAI and Anthropic now

its almost as good as 5.2 and 4.5 but way faster and cheaper

mmaunder 12/17/2025||
Used the hell out of Gemini 3 Flash with some 3 Pro thrown in for the past 3 hours on CUDA/Rust/FFT code that is performance critical, and now have a gemini flavored cocaine hangover and have gone crawling back to Codex GPT 5.2 xhigh and am making slower progress but with higher quality code.

Firstly, 3 Flash is wicked fast and seems to be very smart for a low latency model, and it's a rush just watching it work. Much like the YOLO mode that exists in Gemini CLI, Flash 3 seems to YOLO into solutions without fully understanding all the angles e.g. why something was intentionally designed in a way that at first glance may look wrong, but ended up this way through hard won experience. Codex gpt 5.2 xhigh on the other hand does consider more angles.

It's a hard come-down off the high of using it for the first time because I really really really want these models to go that fast, and to have that much context window. But it ain't there. And turns out for my purposes the longer chain of thought that codex gpt 5.2 xhigh seems to engage in is a more effective approach in terms of outcomes.

And I hate that reality because having to break a lift into 9 stages instead of just doing it in a single wicked fast run is just not as much fun!

JeremyHerrman 12/17/2025||
Disappointed to see continued increased pricing for 3 Flash (up from $0.30/$2.50 to $0.50/$3.00 for 1M input/output tokens).

I'm more excited to see 3 Flash Lite. Gemini 2.5 Flash Lite needs a lot more steering than regular 2.5 Flash, but it is a very capable model and combined with the 50% batch mode discount it is CHEAP ($0.05/$0.20).

jeppebemad 12/17/2025|
Have you seen any indications that there will be a Lite version?
summerlight 12/17/2025||
I guess if they want to eventually deprecate the 2.5 family they will need to provide a substitute. And there are huge demands for cheap models.
heliophobicdude 12/17/2025||
Any word on if this using their diffusion architecture?
aoeusnth1 12/18/2025|
tokens/s don't match, so unlikely
hereme888 12/18/2025||
Any word on when fine-tuning might become available?
More comments...