logoalt Hacker News

NitpickLawyeryesterday at 6:52 PM2 repliesview on HN

> edit: oh small world the cited comment was actually a response to you in that other thread :D

That's hilarious, we must have the same interests since we keep cross posting :D

The thing with the go comparison is that alphago was meant to solve go and nothing else. It couldn't do chess with the same weights.

The current SotA LLMs are "unreasonably good" at a LOT of tasks, while being trained with a very "simple" objective: NTP. That's the key difference here. We have these "stochastic parrots" + RL + compute that basically solve top tier competitions in math, coding, and who knows what else... I think it's insanely good for what it is.


Replies

tech_kenyesterday at 7:00 PM

> I think it's insanely good for what it is.

Oh totally! I think that the progress made in NLP, as well as the surprising collision of NLP with seemingly unrelated spaces (like ICPC word problems) is nothing sort of revolutionary. Nevertheless I also see stuff like this: https://dynomight.substack.com/p/chess

To me this suggests that this out-of-domain performance is more like an unexpected boon, rather than a guarantee of future performance. The "and who knows what else..." is kind of I'm getting: so far we are turning out to be bad at predicting where these tools are going to excel or fall short. To me this is sort of where the "wall" stuff comes from; despite all the incredible successes in these structured problem domains, nobody (in my personal opinion) has really unlocked the "killer app" yet. My belief is that by accepting their limitations we might better position ourselves to laser-target LLMs at the kind of things they rule at, rather than trying to make them "everything tools".

tempusalariayesterday at 7:14 PM

A lot of the current code and science capabilities do not come from NTP training.

Indeed in seems in most language model RL there is not even process supervision, so a long way from NTP