Seems like a hassle when open source models are just as good. Can go with any hosting provider. Might have to wait 3-4 weeks for them to duplicate whatever Anthropic is doing with token caching. But then you get 10x cheaper inference.
I feel like this game is just a hot potato, can you get retail to hold the bag game
Open models are very far in performance from the top models of Anthropic, OpenAI and Google. And that's skipping over the fact you need somewhere to host them.
I have them all. They're not just as good. Whoever tells you that looked only at the benchmarks, not real use. They all fall short at some point.
Kimi K2.5 is the best one, but it's still not at the level of what Anthropic released with opus 4.5.