Top
Best
New

Posted by ingve 9 hours ago

Raspberry Pi's New AI Hat Adds 8GB of RAM for Local LLMs(www.jeffgeerling.com)
193 points | 146 commentspage 3
wyldfire 6 hours ago||
I wonder -- how does this thing compare against the Rubik Pi [1]?

[1] https://rubikpi.ai/

philipallstar 6 hours ago|
Is that affiliated with raspberry pis in some way, or are they just freeloading on the "pi" suffix to confuse people?
wyldfire 3 hours ago|||
They are "freeloading" indeed but in this case it is a valuable indicator that it should be physically compatible with many add-ons/attachments that are compatible with the Raspberry Pi.
Elfener 6 hours ago|||
> are they just freeloading on the "pi" suffix to confuse people?

Yes, but that is normal I guess:

- https://banana-pi.org/

- http://www.orangepi.org/index.html

- https://radxa.com/products/rockpi

incomingpain 3 hours ago||
I wonder if this is the magic hardware for LiquidAI/LFM2.5-Audio-1.5B

Dont need more than 8gb. It'll be enough power. IT can do audio to audio.

renewiltord 8 hours ago||
What’s the current state of the art in low power wake word and speech to text? Has anyone written a blog post on this?

I was able to run a speech to text on my old Pixel 4 but it’s a bit flaky (the background process loses the audio device occasionally). I just want to take some wake word and then send everything to remote LLM and then get back text that I do TTS on.

geerlingguy 8 hours ago||
Maybe not SOTA but the HA Voice Preview Edition [1] in tandem with a Pi 5 or some similar low-power host for the Piper / Whisper pipeline is pretty good. I don't use it but was able to get an Alexa/Google Home-like experience going with minimal effort.

I was only using it for local Home Assistant tasks, didn't try anything further like retrieving sports scores, managing TODO lists, or anything like that.

[1] https://www.home-assistant.io/voice-pe/

folmar 7 hours ago|||
Wake word is not expensive, you can do it on esp32 https://docs.espressif.com/projects/esp-sr/en/latest/esp32s3... (and then send audio to something more beefy as TTS will be marginal at best).
monocasa 8 hours ago||
Wake word can be tiny. Like 10k weights and can run on an esp32 or similar with plenty of compute to spare.

TinyML is a book that goes through the process of building a wake word model for such constrained environments.

huntercaron 8 hours ago||
Glad Jeff was critical here they need a bit of a wake up call it seems.
esskay 7 hours ago||
What a pointless product to waste time making.
rballpug 7 hours ago||
Catalog TG 211, 1000 Hz.
teekert 7 hours ago||
At this moment my two questions for these things are:

1. Can I run a local LLM that allows me to control Home Assistant with natural language? Some basic stuff like timers, to do/shopping lists etc would be nice etc.

2. Can I run object/person detection on local video streams?

I want some AI stuff, but I want it local.

Looks like the answer for this one is: Meh. It can do point 2, but it's not the best option.

worksonmine 7 hours ago||
1. Probably, but not efficiently. But I'm just guessing I haven't tried local LLMs yet.

2. Has been possible in realtime since the first camera was released and has most likely improved since. I did this years ago on the pi zero and it was surprisingly good.

noodletheworld 6 hours ago||
> Can I run a local LLM that allows me to control Home Assistant with natural language? Some basic stuff like timers, to do/shopping lists etc would be nice etc.

No. Get the larger PI recommended by the article.

Quote from the article:

> So power holds it back, but the 8 gigs of RAM holds back the LLM use case (vs just running on the Pi's CPU) the most. The Pi 5 can be bought in up to a 16 GB configuration. That's as much as you get in decent consumer graphics cards1.

> Because of that, many quantized medium-size models target 10-12 GB of RAM usage (leaving space for context, which eats up another 2+ GB of RAM).

…

> 8 GB of RAM is useful, but it's not quite enough to give this HAT an advantage over just paying for the bigger 16GB Pi with more RAM, which will be more flexible and run models faster.

The model specs shown for this device in the article are small, and not fit for purpose even for the relatively trivial use case you mentioned.

I mean, look, lots of people have lots of opinions about this (many of them wrong); it’s cheap, you can buy one and try… but, look. The OP really gave it a shot, and results were kind of shit. The article is pretty clear.

Don’t bother.

You want a device with more memory to mess around with for what you want to do.

imtringued 6 hours ago||
This looks pretty nice for what it is. However, the RAM is a bit oversized for the vast majority of applications that will run on this, which is giving a misleading impression of what it is useful for.

I once tried to run a segmentation model based on a vision transformer on a PC and that model used somewhere around 1 GB for the parameters and several gigabytes for the KV cache and it was almost entirely compute bound. You couldn't run that type of model on previous AI accelerators because they only supported model sizes in the megabytes range.

moffkalast 8 hours ago|
> The Pi's built-in CPU trounces the Hailo 10H.

Case closed. And that's extremely slow to begin with, the Pi 5 only gets what, a 32 bit bus? Laughable performance for a purpose built ASIC that costs more than the Pi itself.

> In my testing, Hailo's hailo-rpi5-examples were not yet updated for this new HAT, and even if I specified the Hailo 10H manually, model files would not load

Laughable levels of support too.

As another datapoint, I've recently managed to get the 8L working natively on Ubuntu 24 with ROS, but only after significant shenanigans involving recompiling the kernel module and building their library for python 3.12 that Hailo for some reason does not provide outside 3.11. They only support the Pi OS (like anyone would use that in prod) and even that is very spotty. Like, why would you not target the most popular robotics distro for an AI accelerator? Who else is gonna buy these things exactly?

More comments...