logoalt Hacker News

satvikpendemyesterday at 11:10 AM7 repliesview on HN

That is "fundamentally" not true, you can use a preset seed and temperature and get a deterministic output.


Replies

HPsquaredyesterday at 11:19 AM

I'll grant that you can guarantee the length of the output and, being a computer program, it's possible (though not always in practice) to rerun and get the same result each time, but that's not guaranteeing anything about said output.

show 2 replies
mhitzayesterday at 12:48 PM

If you self-host an LLM you'll learn quickly that even batching, and caching can affect determinism. I've ran mostly self-hosted models with temp 0 and seen these deviations.

phlakatonyesterday at 2:37 PM

But you cannot predict a priori what that deterministic output will be – and in a real-life situation you will not be operating in deterministic conditions.

zbentleyyesterday at 12:13 PM

Practically, the performance loss of making it truly repeatable (which takes parallelism reduction or coordination overhead, not just temperature and randomizer control) is unacceptable to most people.

show 1 reply
4ndrewlyesterday at 11:41 AM

If you also control the model.

simianparrotyesterday at 11:21 AM

A single byte change in the input changes the output. The sentence "Please do this for me" and "Please, do this for me" can lead to completely distinct output.

Given this, you can't treat it as deterministic even with temp 0 and fixed seed and no memory.

show 4 replies
yunohnyesterday at 11:21 AM

I initially thought the same, but apparently with the inaccuracies inherent to floating-point arithmetic and various other such accuracy leakage, it’s not true!

https://arxiv.org/html/2408.04667v5

show 1 reply