logoalt Hacker News

llagerloftoday at 10:52 AM3 repliesview on HN

What’s new regarding llmfit?

https://github.com/AlexsJones/llmfit


Replies

karmakazetoday at 1:19 PM

This has a web version[0] which I wish they'd host on a free site.

[0] https://github.com/AlexsJones/llmfit/tree/main/llmfit-web

Edit: I tried to deploy a snapshot of the llmfit-web files on Netlify but it seems to need/want to talk to a backend[1]

[1] https://llmfit.netlify.app/

rvztoday at 10:55 AM

Other than it (whichllm) being written in Python, nothing else.

I just use llmfit.

andyyyy64today at 12:44 PM

Fair question. llmfit answers "will this model fit in my memory?" — it's a fit/size calculator, and a good one. whichllm answers a different question: "of the models that fit, which is actually best?" It pulls candidates, then ranks them by merged real benchmarks (LiveBench / Artificial Analysis / Aider / Arena ELO / Open LLM Leaderboard) with a recency penalty, so a newer 27B beats an older 32B even though both fit — on a 24GB card it puts Qwen3.6-27B above Qwen3-32B on benchmarks, not size.

If "biggest that fits" is the answer you want, llmfit is the simpler tool and Python won't matter to you. If you want "which fitting model is worth running," that ranking layer is the whole reason whichllm exists. Different jobs — I'd genuinely send fit-only users to llmfit.

show 4 replies