logoalt Hacker News

0xbadcafebeetoday at 6:30 AM3 repliesview on HN

No mention of the fact that Ollama is about 1000x easier to use. Llama.cpp is a great project, but it's also one of the least user friendly pieces of software I've used. I don't think anyone in the project cares about normal users.

I started with Ollama, and it was great. But I moved to llama.cpp to have more up-to-date fixes. I still use Ollama to pull and list my models because it's so easy. I then built my own set of scripts to populate a separate cache directory of hardlinks so llama-swap can load the gguf's into llama.cpp.


Replies

AndroTuxtoday at 7:00 AM

Exactly. The blog post states that the alternatives listed are similarly intuitive. They are not. If you just need a chat app, then sure, there’s plenty of options. But if you want an OpenAI compatible API with model management, accessibility breaks down fast.

I’m open to suggestions, but the alternatives outlined in the blog post ain’t it.

show 3 replies
BrissyCodertoday at 7:18 AM

> No mention of the fact that Ollama is about 1000x easier to use.

Easier than what?

I came across LM Studio (mentioned in the post) about 3 years ago before I even knew what Ollama as. It was far better even then.

throw9393rjtoday at 7:34 AM

I spend like 2 hours trying to get vulkan acceleration working with ollama, no luck (half models are not supported and crash it). With llama.cpp podman container starts and works in 5 minutes.