logoalt Hacker News

abhikul0yesterday at 3:09 PM1 replyview on HN

I'll try to use that, but llama-server has mmap on by default and the model still takes up the size of the model in RAM, not sure what's going on.


Replies

zozbot234yesterday at 3:14 PM

Try running CPU-only inference to troubleshoot that. GPU layers will likely just ignore mmap.