logoalt Hacker News

tom_0last Saturday at 10:34 AM1 replyview on HN

GGML still runs on llama.cpp, and that still requires CUDA to be installed, unfortunately. I saw a PR for DirectML, but I'm not really holding my breath.


Replies

lostmsuyesterday at 12:37 AM

You don't have to install the whole CUDA. They have a redistributable.

show 1 reply