Grey market fast-follow via distillation seems like an inevitable feature of the near to medium future.
I've previously doubted that the N-1 or N-2 open weight models will ever be attractive to end users, especially power users. But it now seems that user preferences will be yet another saturated benchmark, that even the N-2 models will fully satisfy.
Heck, even my own preferences may be getting saturated already. Opus 4.5 was a very legible jump from 4.1. But 4.6? Apparently better, but it hasn't changed my workflows or the types of problems / questions I put to it.
It's poetic - the greatest theft in human history followed by the greatest comeuppance.
No end-user on planet earth will suffer a single qualm at the notion that their bargain-basement Chinese AI provider 'stole' from American big tech.
not allowing distillation should be illegal :)
One can create 1000s of topic specific AI generated content websites, as a disclaimer each post should include prompt and used model.
Others can "accidentally" crawl those websites and include in their training/fine-tuning.
Just to say - 4.6 really shines on working longer without input. It feels to me like it gets twice as far. I would not want to go back.
In some ways, Opus 4.6 is a step backwards due to massively higher token consumption.
I have no idea how an LLM company can make any argument that their use of content to train the models is allowed that doesn't equally apply to the distillers using an LLM output.
"The distilled LLM isn't stealing the content from the 'parent' LLM, it is learning from the content just as a human would, surely that can't be illegal!"...