logoalt Hacker News

prodigycorpyesterday at 12:08 PM2 repliesview on HN

Are you sure about that? Chain of thought does not need to be semantically useful to improve LLM performance. https://arxiv.org/abs/2404.15758


Replies

davidguettayesterday at 12:11 PM

still doesn't mean all tokens are useful. it's the point of benchmarks

show 1 reply
kubbyesterday at 6:08 PM

If you're misusing LLMs to solve TC^0 problems, which is what the paper is about, then... you also don't need the slop lavine. You can just inject a bunch of filler tokens yourself.