logoalt Hacker News

tcdenttoday at 4:09 AM1 replyview on HN

Furthermore, all of the major LLM APIs reward you for re-sending the same context with only appended data in the form of lower token costs (caching).

There may be a day when we retroactively edit context, but the system in it's current state is not very supportive of that.


Replies

vanviegentoday at 7:17 AM

> Furthermore, all of the major LLM APIs reward you for re-sending the same context with only appended data in the form of lower token costs (caching).

There's a little more flexibility than that. You can strip of some trailing context before appending some new context. This allows you to keep the 'long-term context' minimal, while still making good use of the cache.