logoalt Hacker News

dist-epochtoday at 7:35 AM0 repliesview on HN

For those wondering where is this practically relevant - this is the basic metric used to compare quantization of various LLM models - what is the KL divergence of a 4-bit quantization versus an 8 bit one versus the original 16 bit one.