Top
Best
New

Posted by jeffmcjunkin 11 hours ago

Google releases Gemma 4 open models(deepmind.google)
1161 points | 346 commentspage 6
virgildotcodes 9 hours ago|
Downloaded through LM Studio on an M1 Max 32GB, 26B A4B Q4_K_M

First message:

https://i.postimg.cc/yNZzmGMM/Screenshot-2026-04-03-at-12-44...

Not sure if I'm doing something wrong?

This more or less reflects my experience with most local models over the last couple years (although admittedly most aren't anywhere near this bad). People keep saying they're useful and yet I can't get them to be consistently useful at all.

solarkraft 9 hours ago||
Wow, just like its larger brother!

I had a similarly bad experience running Qwen 3.5 35b a3b directly through llama.cpp. It would massively overthink every request. Somehow in OpenCode it just worked.

I think it comes down to temperature and such (see daniel‘s post), but I haven’t messed with it enough to be sure.

flux3125 9 hours ago||
You're not doing anything wrong, that's expected
gunalx 7 hours ago||
We didnt get deepseek v4, but gemma 4. Cant complain.
DeepYogurt 9 hours ago||
maybe a dumb question but what what does the "it" stand for in the 31B-it vs 31B?
bigyabai 9 hours ago|
Instruction Tuned. It indicates that thinking tokens (eg <think> </think>) are not included in training.
flux3125 9 hours ago|||
That’s not what it means. "-it" just indicates the model is instruction-tuned, i.e. trained to follow prompts and behave like an assistant. It doesn’t imply anything about whether thinking tokens like <think>....</think> were included or excluded during training. Thats a separate design choice and varies by model.
DeepYogurt 8 hours ago||
What does that mean for a user of the model? Is the "-it" version more direct with solutions or something?
petu 6 hours ago|||
It means that model was tuned to to act as chat bot. So write a reply on behalf of assistant and stop generating (by inserting special "end of turn" token to signal inference engine to stop generation).

Base model (without instruction/chat tuning) just generates text non stop ("autocomplete on steroids") and text is not necessarily even formatted as chat -- most text in training data isn't dialogue, after all.

nolist_policy 8 hours ago|||
Use the it versions. The other versions are base models without post-training. E.g. base models are trained to regurgitate raw wikipedia, books, etc. Then these base models are post-trained into instruction-tuned models where they learn to act as a chat assistant.
vigneshj 3 hours ago||
Great one to have
daveguy 8 hours ago||
Fyi, it took me a while to find the meaning of the "-it" in some models. That's how Google designates "instruction tuned". Come on Google. Definite your acronyms.
matt765 9 hours ago||
I'll wait for the next iteration
einpoklum 9 hours ago||
D: Di Gi Charat does not like this nyo! Gemma is supposed to help Dejiko-chan nyo!

G: They offered a very compelling benefits package gemma!

heraldgeezer 10 hours ago||
Gemma vs Gemini?

I am only a casual AI chatbot user, I use what gives me the most and best free limits and versions.

daemonologist 10 hours ago||
Gemma will give you the most, Gemini will give you the best. The former is much smaller and therefore cheaper to run, but less capable.

Although I'm not sure whether Gemma will be available even in aistudio - they took the last one down after people got it to say/do questionable stuff. It's very much intended for self-hosting.

BoorishBears 8 hours ago||
Well specifically a congressperson got it to hallucinate stuff about them then wrote an agry letter

But I checked and it's there... but in the UI web search can't be disabled (presumably to avoid another egg on face situation)

worldsavior 10 hours ago||
Gemma is only 10s of billion parameters, Gemini is 100s.
bertili 10 hours ago||
Qwen: Hold my beer

https://news.ycombinator.com/item?id=47615002

xfalcox 10 hours ago||
Comparing a model you can downloads weights for with an API-only model doesn't make much sense.
regularfry 10 hours ago||
My money's on whatever models qwen does release edging ahead. Probably not by much, but I reckon they'll be better coders just because that's where qwen's edge over gemma has always been. Plus after having seen this land they'll probably tack on a couple of epochs just to be sure.
svachalek 10 hours ago||
The Qwen Plus models should be compared to Gemini, not Gemma.
bibimsz 7 hours ago|
is it good? what's it good for?
More comments...