logoalt Hacker News

DiabloD3last Sunday at 11:20 AM2 repliesview on HN

Largely Vulkan. Microsoft internally is a huge consumer of DirectML for specifically the LLM team doing Phi and the Copilot deployment that lives at Azure.


Replies

1gn15last Monday at 11:56 AM

I'm not sure if it's just the implementation, but I tried using llama.cpp on Vulkan and it is much slower than using it on CUDA.

show 1 reply
bl0blast Monday at 3:01 AM

Such a huge consumer that they deprecated it