logoalt Hacker News

bigyabaiyesterday at 5:13 PM1 replyview on HN

If "small models" is the bar, then you can run inference for ~$50 on Raspberry Pi like hardware. I do that with 1.8b-4b models.


Replies

aegis_camerayesterday at 5:15 PM

LFM 450M for vision task, QWEN 9B Q4 for Orchestration, this provides a good result.

show 1 reply