logoalt Hacker News

cjbgkaghyesterday at 3:29 PM1 replyview on HN

There is nothing intrinsic to LLM prevents reproducibility. You can run them deterministically without adding noise, it would just be a lot slower to have a deterministic order of operations, which takes an already bad idea and makes it worse.


Replies

candiddevmikeyesterday at 3:31 PM

Please tell me how to do this with any of the inference providers or a tool like llama.cpp, and make it work across machines/GPUs. I think you could maybe get close to deterministic output, but you'll always risk having some level of randomness in the output.

show 2 replies