logoalt Hacker News

deivid01/22/20251 replyview on HN

What was the model? What kind of performance did you get out of it?

Could you share a link to your project, if it is public?


Replies

JLCarveth01/22/2025

https://github.com/JLCarveth/nutrition-llama

I've had good speed / reliability with TheBloke/rocket-3B-GGUF on Huggingface, the Q2_K model. I'm sure there are better models out there now, though.

It takes ~8-10 seconds to process an image on my M2 Macbook, so not quite quick enough to run on phones yet, but the accuracy of the output has been quite good.