logoalt Hacker News

cwoolfelast Wednesday at 5:46 PM1 replyview on HN

How can I use ByteShape to run LLMs faster on my 32GB MacBook M1 Max? Or has Ollama already optimized that?


Replies

nunodonatolast Wednesday at 6:05 PM

don't use ollama. llama.cpp is better because ollama has an outdated llama.cpp