The problem you describe is real, but I think it can be addressed by improving tooling without any improvement in available LLM technology.
How? Are you thinking of adversarial AI reviewers, runtime tests (also by AI), or something else?
Guess I just don't see how you can take the human out of the loop and replace them with non-deterministic AIs and informal prompts / specs.
How? Are you thinking of adversarial AI reviewers, runtime tests (also by AI), or something else?
Guess I just don't see how you can take the human out of the loop and replace them with non-deterministic AIs and informal prompts / specs.