logoalt Hacker News

feorentoday at 2:46 PM1 replyview on HN

The final output of the neural network part of an LLM is a vector with weights for every token, that is then usually softmaxed and picked from. Can we not quantify the uncertainty by looking at the distribution of weights of the top 10 options? Like we expect for a note-taking app that the top choice would be something like 98% certain, and if we see that the model gives a weight of 60% to "Russia" and 30% to "France", that's just not enough certainty to simply output "Russia". That's exactly when it should say "<uncertain>" or something instead.


Replies

cyanydeeztoday at 6:24 PM

Unfortunately, that likely just doesn't exist. Everything suggests that these models are confident about their mistakes.