logoalt Hacker News

make3yesterday at 10:17 PM0 repliesview on HN

There's a million algorithms to make LLM inference more efficient as a tradeoff for performance, like using a smaller model, using quantized models, using speculative decoding with a more permissive rejection threshold, etc etc