Posted by jeffmcjunkin 6 hours ago
https://unsloth.ai/docs/models/gemma-4 > Gemma 4 GGUFs > "Use this model" > llama.cpp > llama-server -hf unsloth/gemma-4-31B-it-GGUF:Q8_0
If you already have llama.cpp you might need to update it to support Gemma 4.
Gemma 3 was the first model that I have liked enough to use a lot just for daily questions on my 32G gpu.
How does the ecosystem work? Have things converged and standardized enough where it's "easy" (lol, with tooling) to swap out parts such as weights to fit your needs? Do you need to autogen new custom kernels to fix said things? Super cool stuff.
- Lattner tweeted a link to this: https://www.modular.com/blog/day-zero-launch-fastest-perform...
- Unsloth prior post on gemma 3 finetuning: https://unsloth.ai/blog/gemma3
# with uvx
uvx litert-lm run \
--from-huggingface-repo=litert-community/gemma-4-E2B-it-litert-lm \
gemma-4-E2B-it.litertlmSeems like Google and Anthropic (which I consider leaders) would rather keep their secret sauce to themselves – understandable.
One more thing about Google is that they have everything that others do not:
1. Huge data, audio, video, geospatial 2. Tons of expertise. Attention all you need was born there. 3. Libraries that they wrote. 4. Their own data centers and cloud. 4. Most of all, their own hardware TPUs that no one has.
Therefore once the bubble bursts, the only player standing tall and above all would be Google.
Maybe the model is good but the product is so shitty that I can't perceive its virtues while using it. I would characterize it as pretty much unusable (including as the "Google Assistant" on my phone).
It's extremely frustrating every way that I've used it but it seems like Gemini and Gemma get nothing but praise here.
ChatGPT got it right first time. Baffling.
The other thing that kills me about Gemini is that the voice recognition is god-awful. All of the chat interfaces I use have transcriptions that include errors (which the bot usually treats unthinkingly as what I actually said, instead of acting as if we may be using a fallible voice transcription), but Gemini's is the worst by far. I often have to start conversations over because of such badly mangled transcriptions.
The accuracy problems are the biggest and most important frustrations, but I also find Gemini insufferably chummy and condescending. It often resorts to ELI5 metaphors when describing things to me where the whole metaphor is based on some tenuous link to some small factoid it thinks it remembers about my life.
The experiences it seems people get out of Gemini today seem like a waste of a frontier lab's resources tbf. If I wanted fast but lower quality I'd go to one of the many smaller providers that aren't frontier labs because lots of them are great at speed and/or efficiency. (If I wanted an AI companion, Google doesn't seem like the right choice either.)
Really eager to test this version with all the extra capabilities provided.
Others have just borrowed data, money, hardware and they would run out of resources for sure.