logoalt Hacker News

catlifeonmarsyesterday at 6:13 AM1 replyview on HN

It seems kind of silly that you can’t teach an LLM new tricks though, doesn’t it? This doesn’t sound like an intrinsic limitation and more an artifact of how we produce model weights today.


Replies

verdvermyesterday at 1:46 PM

getting tricks embedded into the weights is expensive, it doesn't happen in a single pass

they's why we teach them new tricks on the fly (in-context learning) with instruction files

show 1 reply