I still haven't seen any statistically sound data supporting that this is happening on the API (per-token pricing.)
If you've got something to share I'd love to see it.
There's an interesting analysis here: https://github.com/anthropics/claude-code/issues/42796
>The most striking row is user prompts: 5,608 in February vs 5,701 in March. The human put in the same effort. But the model consumed 80x more API requests and 64x more output tokens to produce demonstrably worse results.
There's an interesting analysis here: https://github.com/anthropics/claude-code/issues/42796
>The most striking row is user prompts: 5,608 in February vs 5,701 in March. The human put in the same effort. But the model consumed 80x more API requests and 64x more output tokens to produce demonstrably worse results.