logoalt Hacker News

cr125rideryesterday at 6:17 PM2 repliesview on HN

Seems like it’s maybe also a tool steering problem. These models should be reaching for tools to help solve factual problems. LLM should stick to prose.


Replies

emp17344yesterday at 6:23 PM

I think this is still useful research that calls into question how “smart” these models are. If the model needs a separate tool to solve a problem, has the model really solved the problem, or just outsourced it to a harness that it’s been trained - via reinforcement learning - to call upon?

show 2 replies
stratos123yesterday at 7:03 PM

Are you still talking about this paper? No tools were allowed in it.