logoalt Hacker News

kgeistyesterday at 2:28 PM0 repliesview on HN

Llama.cpp already uses an idea from it internally for the KV cache [0]

So a quantized KV cache now must see less degradation

[0] https://github.com/ggml-org/llama.cpp/pull/21038