-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Open
Description
I'm experimenting with Llamafile on a Raspberry Pi 5 with 8Gb of ram, in order to integrate it with existing privacy-protecting smart home voice control. This is working great so far as long as I very small models are used.
I was wondering: would it be possible to speed up inference on the Rasbperry Pi 5 by using the GPU?
Through this Stack Overflow post I've found some frameworks that already do this, such as:
- NCNN https://github.com/Tencent/ncnn/wiki/how-to-build#build-for-linux
- BeatMUp https://github.com/lnstadrum/beatmup/
The Raspberry Pi 5's VideoCore GPU has vulkan drivers:
https://www.phoronix.com/news/Mesa-RPi-5-VideoCore-7.1.x
Curious to your thoughts.
Related:
#40
mounta11n