Top
Best
New

Posted by AbuAssar 6 days ago

Lemonade by AMD: a fast and open source local LLM server using GPU and NPU(lemonade-server.ai)
570 points | 112 commentspage 3
Sparkyte 5 days ago|
What is the lowest process I can implement this on?
LowLevelKernel 5 days ago||
Which specific NPU’s?
robotswantdata 5 days ago||
Forget all the vibe coded slop or Ollama. Lemonade is the real deal and very good, been using about a year now.

AMD are doing gods work here

ozgrakkurt 5 days ago||
For people with AMD card. This is garbage, rocm is garbage. Just install llama.cpp and run llama-server with vulkan option. This is just some slop + JS/Electron garbage put on top.
9dc 6 days ago||
so... what does it do? i dont get it Lol
iugtmkbdfil834 6 days ago|
Initial read suggests it is a mini-swiss army knife, because it seems to be able to do a lot ( based on website claims anyway ). The app integration seems to suggest they want to be more of a control dashboard.
techpulselab 5 days ago||
[flagged]
johnwhitman 4 days ago||
[flagged]
philbitt 5 days ago||
[dead]
devnotes77 5 days ago||
[dead]
aplomb1026 5 days ago|
[dead]
More comments...