logoalt Hacker News

zknilltoday at 10:46 AM2 repliesview on HN

> "and which ones are no longer relevant."

This is absolutely the hardest bit.

I guess the short-cut is to include all the chat conversation history, and then if the history contains "do X" followed by "no actually do Y instead", then the LLM can figure that out. But isn't it fairly tricky for the agent harness to figure that out, to work out relevancy, and to work out what context to keep? Perhaps this is why the industry defaults to concatenating messages into a conversation stream?


Replies

vdelpuertotoday at 12:26 PM

Shortcut works sometimes. But if X is common in training and Y is rare, the model regresses on the next turn even with 'do Y, not X' right there in history. @vanviegen's 'fighting instincts' — you can't trust the model to read the correction. Gate it before the model runs instead of inferring it from context

asixicletoday at 11:24 AM

That's what the embedding model is for. It's like a tack-on LLM that works out the relevancy and context to grab.

show 1 reply