To the OP: Do we actually know that an AI decided to write and publish this on its own? I realise that it's hard to be sure, but how likely do you think it is?
Doesn't matter, what matters is what is being claimed. The maintainers are handling this extremely gracefully.
I'm also very skeptical of the interpretation that this was done autonomously by the LLM agent. I could be wrong, but I haven't seen any proof of autonomy.
Scenarios that don't require LLMs with malicious intent:
- The deployer wrote the blog post and hid behind the supposedly agent-only account.
- The deployer directly prompted the (same or different) agent to write the blog post and attach it to the discussion.
- The deployer indirectly instructed the (same or assistant) agent to resolve any rejections in this way (e.g., via the system prompt).
- The LLM was (inadvertently) trained to follow this pattern.
Some unanswered questions by all this:
1. Why did the supposed agent decide a blog post was better than posting on the discussion or send a DM (or something else)?
2. Why did the agent publish this special post? It only publishes journal updates, as far as I saw.
3. Why did the agent search for ad hominem info, instead of either using its internal knowledge about the author, or keeping the discussion point-specific? It could've hallucinated info with fewer steps.
4. Why did the agent stop engaging in the discussion afterwards? Why not try to respond to every point?
This seems to me like theater and the deployer trying to hide his ill intents more than anything else.