It's so easy to google this and find that they all do exactly this.
Gemini: https://ai.google.dev/gemini-api/docs/code-execution
ChatGPT: https://help.openai.com/en/articles/8437071-data-analysis-wi...
Claude: https://claude.com/blog/analysis-tool
Reasoning only gets you so far, even humans write code or use spreadsheets, calculators, etc, to get their answers to problems.
you have just linked the fact that they have code executions but not proved that it is needed for strawberry problem.
there are multiple ways to disprove this
1. GPT o1 was released and it never supported the tools and it easily solved the strawberry problem - it was named strawberry internally
2. you can run GPT 5.2-thinking in the API right now and deny access to any tools, it will still work
3. you can run deepseek locally without tools and run it, it will still work
Overall this idea that LLM's cant reason and need tools to do that is misleading and false and easily disproven.