logoalt Hacker News

Something is afoot in the land of Qwen

424 pointsby simonwtoday at 3:55 PM206 commentsview on HN

Comments

sosodevtoday at 4:22 PM

I really hope this doesn't hinder development too much. As Simon says, Qwen3.5 is very impressive.

I've been testing Qwen3.5-35B-A3B over the past couple of days and it's a very impressive model. It's the most capable agentic coding model I've tested at that size by far. I've had it writing Rust and Elixir via the Pi harness and found that it's very capable of handling well defined tasks with minimal steering from me. I tell it to write tests and it writes sane ones ensuring they pass without cheating. It handles the loop of responding to test and compiler errors while pushing towards its goal very well.

show 8 replies
hintymadtoday at 5:54 PM

There has been tension between Qwen's research team and Alibaba's product team, say the Qwen App. And recently, Alibaba tried to impose DAU as a KPI. It's understandable that a company like Alibaba would force a change of product strategy for any number of reasons. What puzzled me is why they would push out the key members of their research team. Didn't the industry have a shortage of model researchers and builders?

show 1 reply
softwaredougtoday at 4:24 PM

I wonder how a US lab hasn't dumped truckloads of cash into various laps to ensure these researchers have a place at their lab

show 9 replies
lzaborowskitoday at 7:50 PM

One thing I’ve noticed with local models is that people tolerate a lot more trial and error behavior. When a hosted model wastes tokens it feels expensive, but when a local model loops a bit it just feels like it’s “thinking.”

If models like Qwen can get good enough for coding tasks locally, the real shift might be economic rather than purely capability.

show 1 reply
vicchenaitoday at 8:51 PM

Been running the 32B locally for a few days and honestly surprised how well it handles agentic coding stuff. Definitely punches above its weight. Only complaint is it sometimes decides to ignore half your prompt when instructions get long, but at this size I guess thats the tradeoff.

airstriketoday at 4:10 PM

I'm hopeful they will pick up their work elsewhere and continue on this great fight for competitive open weight models.

To be honest, it's sort of what I expected governments to be funding right now, but I suppose Chinese companies are a close second.

skeeter2020today at 4:24 PM

Getting a bit of whiplash goin from AI is replacing people, to AI is dead without (these specific) people. Surely we're far enough ahead that AI can take it from here?

Wild times!

show 4 replies
quantum_statetoday at 5:14 PM

I would second that Qwen3.5 is exceptionally good. In a calibration, it (35b variant) was running locally with Ada NextGen 24GB to do the same things with easy-llm-cli in comparison with gemini-cli + Gemini 3 Pro, they were at par … really impressive it ran pretty fast …

show 1 reply
zobatoday at 4:16 PM

I tried the new qwen model in Codex CLI and in Roo Code and I found it to be pretty bad. For instance I told it I wanted a new vite app and it just started writing all the files from scratch (which didn’t work) rather than using the vite CLI tool.

Is there a better agentic coding harness people are using for these models? Based on my experience I can definitely believe the claims that these models are overfit to Evals and not broadly capable.

show 5 replies
lacooljtoday at 7:26 PM

I wonder if an american company poached one/all of them. They've been pretty much bleeding edge of open models and would not surprise me if Amazon or Google snatched them up

show 1 reply
ilakshtoday at 5:00 PM

Does anyone know when the small Qwen 3.5 models are going to be on OpenRouter?

show 1 reply
nurettintoday at 7:09 PM

I am singularly impressed by 35B/A3, hope that is not the reason he had to leave.

raffael_detoday at 4:08 PM

> me stepping down. bye my beloved qwen.

the qwen is dead, long live the qwen.

w10-1today at 8:05 PM

It sounds like the lead was demoted to attract new talent, quit as a result, and the rest of the team also resigned to force management to change their minds.

If so, I'm happy that the team held together, and I hope that endogenous tech leads get to control their own career and tech destiny after hard work leads to great products. (It's almost as inspiring as tank man, and the tank commanders who tried to avoid harming him...)

(ducking the downvote for challenging the primacy of equity...)

hwerstoday at 4:44 PM

My conspiracy theory hat is that somehow investors with a stake in openai as well is sabotaging, like they did when kicking emad out of stabilityai

show 2 replies
vonneumannstantoday at 4:32 PM

Were they kneecapped by Anthropic blocking their distillation attempts?

show 1 reply
kartika848484today at 6:29 PM

what the hell, their models were promising tho

aplomb1026today at 6:31 PM

[dead]

butILoveLifetoday at 4:36 PM

[flagged]

show 4 replies
multisporttoday at 4:36 PM

inb4 qwen is less of a supply chain risk than anthropic