Top
Best
New

Posted by lairv 10 hours ago

Ggml.ai joins Hugging Face to ensure the long-term progress of Local AI(github.com)
608 points | 143 commentspage 2
karmasimida 3 hours ago|
Does local AI have a future? The models are getting ridiculously big and any storage hardware is hoarded by few companies for next 2 years and nvidia has stopped making consumer GPU for this year.

It seems to me there is no chance local ML is going to be anywhere out of the toy status comparing to closed source ones in short term

rhdunn 3 hours ago||
Mistral have small variants (3B, 8B, 14B, etc.), as do others like IBM Granite and Qwen. Then there are finetunes based on these models, depending on your workflow/requirements.
karmasimida 49 minutes ago||
True, but anything remotely useful is 300B and above
dust42 3 hours ago||
I am actually doing now a good part of dev with Qwen3-Coder-Next on an M1 64GB with Qwen Code CLI (a fork of Gemini CLI). I very much like

  a) to have an idea how much tokens I use and 
  b) be independent of VC financed token machines and 
  c) I can use it on a plane/train
Also I never have to wait in a queue, nor will I be told to wait for a few hours. And I get many answers in a second.

I don't do full vibe coding with a dozen agents though. I read all the code it produces and guide it where necessary.

Last not least, at some point the VC funded party will be over and when this happens one better knows how to be highly efficient in AI token use.

the__alchemist 9 hours ago||
Does anyone have a good comparison of HuggingFace/Candle to Burn? I am testing them concurrently, and Burn seems to have an easier-to-use API. (And can use Candle as a backend, which is confusing) When I ask on Reddit or Discord channels, people overwhelmingly recommend Burn, but provide no concrete reasons beyond "Candle is more for inference while Burn is training and inference". This doesn't track, as I've done training on Candle. So, if you've used both: Thoughts?
csunoser 7 hours ago|
I have used both (albeit 2 years ago, and things change really fast). At the time, Candle didn't have 2d conv backprop with strides properly implemented. And getting Burn running libtch backend was just a lot simpler.

I did use candle for wasm based inference for teaching purposes - that was reasonably painless and pretty nice.

mattfrommars 5 hours ago||
I don’t know if this warrants a separate thread here but I have to ask…

How can I realistically get involved the AI development space? I feel left out with what’s going on and living in a bubble where AI is forced into by my employer to make use of it (GitHub Copilot), what is a realistic road map to kinda slowly get into AI development, whatever that means

My background is full stack development in Java and React, albeit development is slow.

I’ve only messed with AI on very application side, created a local chat bot for demo purposes to understand what RAG is about to running models locally. But all of this is very superficial and I feel I’m not in the deep with what AI is about. I get I’m too ‘late’ to be on the side of building the next frontier model and makes no sense, what else can I do?

I know Python, next step is maybe do ‘LLM from scratch”? Or I pick up Google machine learning crash course certificate? Or do recently released Nvidia Certification?

I’m open for suggestions

fc417fc802 3 hours ago||
I'm not entirely clear what your goals are but roughly, just figure out an application that holds your interest and build a model for it from scratch. Probably don't start with an LLM though. Same as for anything else really. If you're interest in computer graphics then decide on a small scale project and go build it from scratch. Etc.
breisa 3 hours ago||
Maybe look into model finetuning/distilation. Unsloth [1] has great guides and provides everything you need to get started on Google Colab for free. [1] https://unsloth.ai/
moralestapia 2 hours ago||
I hope Georgi gets a big fat check out of this, he deserves it 100%.
jimmydoe 10 hours ago||
Amazing. I like the openness of both project and really excited for them.

Hopefully this does not mean consolidation due to resource dry up but true fusion of the bests.

forty 2 hours ago||
Looks like someone tried to type "Gmail" while drunk...
rkomorn 2 hours ago|
Looks like Gargamel of Smurfs fame to me.
androiddrew 9 hours ago||
One of the few acquisitions I do support
cyanydeez 2 hours ago||
Is there a local webui that integrates with Hugging face?

Ollama and webui seem to rapidly lose their charm. Ollama now includes cloud apis which makes no sense as a local.

sheepscreek 7 hours ago||
Curious about the financials behind this deal. Did they close above what they raised? What’s in it for HuggingFace?
stephantul 8 hours ago|
Georgi is such a legend. Glad to see this happening
More comments...