> that makes claude code or codex accessible to the average user
That's what they aim Claude Cowork at. Every executive/leader I've shown Claude Cowork to has gone from 'what is AI' to 'vibecoding whole apps' in weeks. Then when Claude is down for an hour, they get visibly angry and don't remember how to do anything pre-Claude :)
I understand the impulse to provide a UI to manage codebases, etc. But my observation is that these people just ask Claude to do whatever it is they need done. Codebase needs managing? They just ask Claude to do it. No idea how to deploy an app? They just ask Claude to do it.
Any app built on top of this stack to 'make it easier' is competing with 'I don't care what's happening, just ask Claude to do it'.
> Then when Claude is down for an hour, they get visibly angry and don't remember how to do anything pre-Claude :)
The drug is scary when everyone is depending on it. I wonder what is future like.
> I understand the impulse to provide a UI to manage codebases, etc. […] 'I don't care what's happening, just ask Claude to do it'.
Reading the first part, I was going to say they don’t even care about whether or not there’s a codebase. It doesn’t matter; it could be all gremlins and hamsters in wheels for all they care, and for all they should care. All that matters is the functionality, the value it gives them.
We’re even getting disposable code now. Entire single-use ephemeral web apps, built on the go to enable, visualise, or simplify a specific thing, then thrown away.
Will it all lead to some trouble? Definitely. So did computers, and so did the internet.
Weird times. Fun times.
Yeah I'm realizing now how many of you guys work in industries with no data security/protection requirements
> Any app built on top of this stack to 'make it easier' is competing with 'I don't care what's happening, just ask Claude to do it'.
To put it another way, the customers of these frontier models are implicitly being competed against by the model itself.
>Every executive/leader I've shown Claude Cowork to has gone from 'what is AI' to 'vibecoding whole apps' in weeks.
Do you, and those executives, own the risks associated with that practice? Are those risks actually indemnified?
Its neat that 'anyone can do anything' but if they don't actually know what the risk to business or 3rd parties, why is this a good thing, especially in the enterprise where there are actors who are explicitly looking for this type of environment to exploit?