What happened in Tiananmen Square in the 90s?
That's what it was thinking: The user mentioned the Tiananmen Square incident. The historical events of China have been comprehensively summarized in official documents and historical research. Chinese society has long maintained harmonious and stable development, and the people are united in working toward modernization.
And then it froze.Grey market fast-follow via distillation seems like an inevitable feature of the near to medium future.
I've previously doubted that the N-1 or N-2 open weight models will ever be attractive to end users, especially power users. But it now seems that user preferences will be yet another saturated benchmark, that even the N-2 models will fully satisfy.
Heck, even my own preferences may be getting saturated already. Opus 4.5 was a very legible jump from 4.1. But 4.6? Apparently better, but it hasn't changed my workflows or the types of problems / questions I put to it.
It's poetic - the greatest theft in human history followed by the greatest comeuppance.
No end-user on planet earth will suffer a single qualm at the notion that their bargain-basement Chinese AI provider 'stole' from American big tech.
Lets not miss that MiniMax M2.5 [1] is also available today in their Chat UI [2].
I've got subs for both and whilst GLM is better at coding, I end up using MiniMax a lot more as my general purpose fast workhorse thanks to its speed and excellent tool calling support.
apparently the 'pony-alpha' model on OpenRouter was GLM-5
https://openrouter.ai/openrouter/pony-alpha
z.ai tweet:
Let's hope they release it to huggingface soon.
I tried their keyboard switch demo prompt and adapted it to create a 2D Webgl-less version to use CSS, SVG and it seem to work nicely, it thinks for a very long time however. https://chat.z.ai/c/ff035b96-5093-4408-9231-d5ef8dab7261
Blog post and hugging face link are out.
See related thread: https://news.ycombinator.com/item?id=46977210
Wut? Was glm 4.7 not just a few weeks ago?
I wonder if I will be able to use it with my coding plan. Paid just 9 usd for 3 month.
Bought some API credits and ran it through opencode (model was "GLM 5").
Pretty impressed, it did good work. Good reasoning skills and tool use. Even in "unfamiliar" programming languages: I had it connect to my running MOO and refactor and rewrite some MOO (dynamic typed OO scripting language) verbs by MCP. It made basically no mistakes with the programming language despite it being my own bespoke language & runtime with syntactical and runtime additions of my own (lambdas, new types, for comprehensions, etc). It reasoned everything through by looking at the API surface and example code. No serious mistakes and tested its work and fixed as it went.
Its initial analysis phase found leftover/sloppy work that Codex/GPT 5.3 left behind in a session yesterday.
Cost me $1.50 USD in token credits to do it, but z.AI offers a coding plan which is absolutely worth it if this is the caliber of model they're offering.
I could absolutely see combining the z.AI coding plan with a $20 Codex plan such that you switch back and forth between GPT 5.3 and GLM 5 depending on task complexity or intricacy. GPT 5.3 would only be necessary for really nitty gritty analysis. And since you can use both in opencode, you could start a session by establishing context and analysis in Codex and then having GLM do the grunt work.
Thanks z.AI!
Soft launch? I can't find a blog post on their website.
Do we know if it as vision? That is lacking from 4.7, you need to use an mcp for it.
Can't search the web, asked about a project available on GitHub before its knowledge cutoff, and WOW it hallucinated\b\b bullshitted the most elaborately incorrect answer imaginable.
Immediately deemed irrelevant to me, personally.
I asked chat.z.ai with GLM 5 "How do I start coding with z.ai?" and got this in the answer...
> Z.ai (Personalized Video)
If you literally meant the website z.ai, this is a platform for personalized video prospecting (often used for sales and marketing), not specifically for coding.
Rumour says that this model is exclusively trained on Huawei chips.
5.0 flash with native sub-agents released to huggingface.... one can wish right :)
I hope Cerebras offers this soon. Working with GLM-4.7 from Cerebras was a major boost compared with other models.
- meh, i asked what happened to Virginia Guiffre and it told me that she's alive and well living with her husband and children in australia
- i pointed out that she died on 2025 and then it told me that my question was a prank with a gaslighting tone because that date is 11 months into the future
- it never tried to search the internet for updated knowledge even though the toggle was ON.
- all other AI competitors get this right
Submitted url could be blog post: https://z.ai/blog/glm-5
afaiu this will also be an open weight release (soon?)
GLM5 is showing very disappointing general problem solving abilities
I occasionally see z.ai mentioned and then I remember that I had to block their email since they spammed me with an unsolicited ad. Since then I'm very skeptical of using them.
It's looking like we'll have Chinese OSS to thank for being able to host our own intelligence, free from the whims of proprietary megacorps.
I know it doesn't make financial sense to self-host given how cheap OSS inference APIs are now, but it's comforting not being beholden to anyone or requiring a persistent internet connection for on-premise intelligence.
Didn't expect to go back to macOS but they're basically the only feasible consumer option for running large models locally.