Is this true? Non-reasoning LLMs are autoregressive. Reasoning LLMs can emit thousands of reasoning tokens before "line 1" where they write the answer.
They are all autoregressive. They have just been trained to emit thinking tokens like any other tokens.
reasoning is just more tokens that come out first wrapped in <thinking></thinking>