I'm having somewhat good experiences with AI but I think that's because I'm only half-adopting it: instead of the full agentic / Ralphing / the-AI-can-do-anything way, I still do work in very small increments and review each commit. I'm not as fast as others, but I can catch issues earlier. I also can see when code is becoming a mess and stop to fix things. I mean, I don't fix them manually, I point Claude at the messy code and ask it to refactor it appropriately, but I do keep an eye to make sure Claude doesn't stray off course.
Honestly, seeing all the dumb code that it produces, calling this thing "intelligent" is rather generous...
I would love it if someone explained what their ten agents Ralphing away were actually told to do.
I suppose if you are doing something that truly can be decided based on a test but, I just don't see it, at least for anything I do.