logoalt Hacker News

bensyversonyesterday at 3:22 PM1 replyview on HN

Exactly. The model is exquisitely sensitive to language. The idea that you would encourage it to think like a caveman to save a few tokens is hilarious but extremely counter-productive if you care about the quality of its reasoning.


Replies

andaiyesterday at 10:21 PM

Does this imply that if you train it on Gwern style output, the quality will improve?

show 1 reply