logoalt Hacker News

jamesharttoday at 1:21 AM0 repliesview on HN

LLMs can have whatever abilities we build for them. The fact we currently start their context out with a static prompt which we keep feeding in on every iteration of the token prediction loop is a choice. We don’t have to keep doing that if there are other options available.