logoalt Hacker News

feisty0630yesterday at 4:58 PM1 replyview on HN

Interesting that it reads a bit like it came from a Markov chain rather than an LLM. Perhaps limited training data?


Replies

kgeistyesterday at 8:54 PM

Early LLMs used to have this often. I think's that where the "repetition penalty" parameter comes from. I suspect output quality can be improved with better sampling parameters.