Holy moly, I made a simple coding promt and the amount of reasoning output could fill a small book.
> create a single html file with a voxel car that drives in a circle.
Compared to GLM 4.7 / 5 and kimi 2.5 it took a while. The output was fast, but because it wrote so I had to wait longer. Also output was .. more bare bones compared to others.
That's been my experience as well. Huge amounts of reasoning. The model itself is good but even if you get twice as many tokens as with another model, the added amount of reasoning may make it slower in the end.
That's how it compensates for its small size. To accomplish a task of certain difficulty either you know more and think less, or vice versa.