Given that providers of open source models can offer Kimi K2.5 at input $0.60 and output $2.50 per million tokens, I think the cost of inference must be around that. We would still need to compare the tokens per second.
Fair but we technically do not know the parameter count
Fair but we technically do not know the parameter count