logoalt Hacker News

bensyversonlast Thursday at 3:22 PM1 replyview on HN

Exactly. The model is exquisitely sensitive to language. The idea that you would encourage it to think like a caveman to save a few tokens is hilarious but extremely counter-productive if you care about the quality of its reasoning.


Replies

andailast Thursday at 10:21 PM

Does this imply that if you train it on Gwern style output, the quality will improve?

show 1 reply