I furthermore wish that "posting an LLM-generated comment (i.e. and passing it off as your own)" was worthy of an instant ban, because I see this sort of behavior from non-green accounts as well.
EDIT: I meant (but totally forgot) to qualify that my "proposal" would only apply when the LLM-ness is self-obvious—idk, make up a "reasonable person" standard or something. Presumably, the moderators would err on the side of letting things slide. Even so, many comments I've seen are simply impossible for any reasonable person to claim as "human-written"—the default ChatGPT style is simply too distinct.
I think your comment was generated by an LLM and hereby vote for your immediate and permanent instant ban.
Many HNers strongly argue that it's absolutely impossible to distinguish between AI text and non-AI text. Some of it seems to be a knee-jerk reaction to some of the occasional, one-sided stories of people who were accused of using LLMs and fired from their jobs. And some of it seems to be just hedging so that we don't develop a culture that could penalize their LLM-generated posts or code.
We had people defending the fired Ars Technica guy, even though he admitted to using an LLM in some sort of a contrived non-apology along the lines of "I did it because I had a cold".
My main problem with that is that you can just generate an infinite supply of LLM op-eds about LLMs, and is this really what we want to read every day? If I want to know what ChatGPT thinks about the risks or benefits of vibecoding, I'll just ask it.
The moderators are supposed to just know it when they see it? It's that black and white to you? Or are lots of false positives a price we have to pay?
People accuses everything of being LLM generated these days. That'd be a tough rule to enforce.
Do this with submissions, too. Or at least put some indicator that it's AI generated.
The guidelines haven't even been updated to say that AI generated posts and submission aren't permitted even though it's been the policy for a couple of years now if one searches for postings by the moderators. So outsiders and new HN users have no reason to know that it's not allowed. I'm sure there are reasons for it but the inaction is all very mysterious from an outsider perspective.
I disagree with this policy.
Some people can really benefit from using LLMs to help them write. E.g. non-native speakers.
LLM-assisted-writing doesn't have to be low effort, it can help people express themselves better in many cases. I'd argue that someone who spent their time doing multiple passes with an LLM to get their phrasing just write, has taken obviously more care than the majority of people on HN take before commenting.
And if you don't like the way something is written? Just down vote it. That's true whether or not it's partially/wholly written by an LLM.
I think all submissions to HN should be submitted via snail-mail, and must be handwritten. That would solve the problem.
/heavy sarcasm
That being said, my mother used to insist on hand-written cover letters from job applicants. Her rationale: it takes effort, so it weeds out all the applications from people who are just randomly spraying out applications for jobs they are not qualified for.
Other than this being probably challenging to enforce fairly, I think I agree that if you had strong proof of an account largely or completely posting comments/stories/whatever that was adulterated by an LLM, that is really probably ban worthy like you said.
When I read comments like this, I think about the average Joe who says: "Most people are terrible drivers." Then, someone asks them: "Are you a terrible driver?" They respond: "Of course not. I am an excellent driver." A few people roll their eyes.
> worthy of an instant ban
First, it is not always possible to identify an LLM-generated comment. There are too many false-positives. Imagine if this system was implemented, and one of your comments was identified as LLM-generated and you were instantly banned. How would you feel about it?How ironic, a comment advocating for banning LLM comments using em dashes
What if someone used an LLM to just translate?
For now there is already a pretty effective mechanism in place, downvote and/or flag those comments that you think are across the line in that sense.
But in principle I agree with you, the rule for me is 'if it wasn't worth your time to write then it certainly isn't worth 1000x times other people's time to read'.
I think you need (at least) one exception to that rule. We have many people here whose first language is not English, and this is an English-only forum. For at least some of those people, an AI translation may give better clarity than their own attempt at writing in English.
So I would propose that, in the ideal world where we could perfectly enforce the rules that we chose, that the rule would be "AI for translation only". If it wrote your content, your comment is gone. If it translated content that you wrote, your comment is still welcome.
> I furthermore wish that "posting an LLM-generated comment (i.e. and passing it off as your own)" was worthy of an instant ban
It pretty much is. It’s not hard and fast (sometimes we’ll warn people or email them to ask if it’s not certain) and it takes time for us to see things and act, especially when people don’t email us when they see these comments.
But as a general rule, accounts that post generated comments get banned.