I have a story with Benji.
Last year I went viral, and Benji was the first person to interview me. It was a really cool experience, we chatted via Twitter dms, and he wrote a piece about my work - overall did a decent job.
Then, 6 months later a separate project I was adjacent to was starting to pick up steam. I reached out to him asking if he wanted to cover us. No response.
Then, tech crunch wrote an article on our project.
I reached to Benji again saying "Hey would you like to chat again, now we have some coverage?" And he finally responded, but said he couldn't report on me because he had a directive that he could only report on things that didn't have any prior or pre-existing coverage (?)
I thought that was rather strange, especially since we already had built up a relationship.
I don't really have a moral or lesson to this story, other than that journalism can be rather opaque sometimes.
Oh one other tip for anyone reading this - if you do ever get reached out to by journalists, communicate in writing, not a phone call so you can be VERY precise in your wordings.
This is exactly why every AI citation we publish goes through a blocker. We dump the AI transcript plus the generated case numbers into a little script that hits the official court database and only passes through citations that return the same case id, party names, and paragraph text. If the extra lookup fails, the shot has to be marked as a hallucination, logged in the docket, and a human has to go re-verify with the actual law reports before we file anything. Treat the LLM like a drafting helper, not an authority, and make the human verification the gate that moves the draft from “AI promised” to “judicially safe.” We also keep a micro audit trail so if a clerk says “the AI gave me this,” we can replay how the prompt went and which citation check failed. What guard rails have other people put in front of AI-written judgements?
The individual firing is a distraction from the structural issue. Newsrooms have been cutting editorial staff for a decade, which means the verification layers that would have caught this — fact-checkers, copy editors, senior editors doing source verification — largely don't exist anymore. Then they adopt AI tools that increase throughput without increasing oversight capacity, and act surprised when fabrication slips through.
This is a classic systems failure: you remove the safety mechanisms, add a new source of risk, and punish the individual operator. It's the same pattern you see in industrial accidents. The Swiss cheese model applies — every editorial layer that got cut was a slice of cheese being removed.
The more interesting policy question is whether publications should be required to disclose AI tool usage in their editorial process, similar to how financial publications disclose conflicts of interest. The FTC has signaled interest in AI-generated content transparency but hasn't issued concrete guidance for journalism yet.
As much as I respect the site and gladly financially support it, this is ultimately a failure on Ars Technica and its editors. If there are any.
If this were just some random blogger, then yes the blame is totally theirs. But this was published under the Ars Technica masthead and there should have been someone or something double checking the veracity of the contents.
That said, there are a number of Ars Technica contributors that are among the best in their fields: Eric Burger, Dan Goodin, Beth Mole, Stephen Clark, and Andrew Cunningham amongst many, so one f'up shouldn't really impugn the entire organization.
I'ts an open secret that even the larger news outlets mandate LLM use. They buy subscriptions and have guidelines on how to mask the output (so that it would read less AI'ed), how to fact-check the links and the quotes etc. The authors which aren't willing to jump on this particular train are quickly let go due to performance.
The expectation is to produce more with much less (staff), the pipeline is heavily optimized for clicks, every single headline is A/B tested- Ars isn't alone in churning out poorly reviewed clickbait (and then not owning their mistakes)
Context from earlier discussion of the article being pulled: https://news.ycombinator.com/item?id=47009949
The headline says Ars fired the reporter, but AFAICT the article doesn't include any facts that indicate this. All we know is that he no longer works there, and that Ars refused to provide any additional information.
The commentariat senses blood in the water and will criticize Ars Technica no matter how they respond here. It seems fine. The author really paid the price. I trust Ars to be extra vigilant to this going forward.
I have to admit, nowadays Google AI Overview's accuracy is so good that I often don't check the links. It's scary that it got from 'practically useless' to 'the actual google search' in less than two years.
I really don't know where the internet is heading to and how any content site can survive.
I don't know that this is what happened here, but any time there is a push to do more with less, you end up rewarding people who take shortcuts over those who do a proper job, and from the outside, it looks like journalism has a push to do more with less.
You will never get the internet to agree on how incident x should have been handled. I think the world right now is running to figure out AI and its place. Just when you think you understand, the ground shifts. It is clear that in the future this exact use of AI will be expected and work, on average, way better than a person. I know that a lot of people probably have an emotional 'no it won't!' and disagree with me here but there have been so many 'no it won't! never!' moments passed in the last two years that I can't imagine this won't also be one. With that in mind I don't think it is reasonable to fire this journalist. They used a tool too soon but it is really hard to figure out what is too soon right now. This should have been a moment of reflection for their news room (and probably some private conversations) but it turned into a firing which I think is too much. Did the news room gain from that? Will it prevent them from doing it again? Did it fix the original mistake? I don't think the answer is 'yes' to any of these questions. A good retraction, apology, statement on how they are changing and will review new technology entering the newsroom in the future. Those help.
These comments are full of people who are not and have never been journalists talking about editorial processes they don't understand. I was a journalist before AI (and have since pivoted to software development), so I feel the need to correct a mistaken assumption.
1. An editor's role is not to make sure all quotes are accurate to what the source spoke or wrote. They are not babysitters. It's a higher-level position that is supposed to ensure that the article makes sense; that the sources are sufficient in number and are credible; that the story covers the assigned topic, etc. Your job as an editor is _not_ to babysit your reporters. Reporters are given a large degree of freedom and are expected to have sufficient education/training and ethical grounding* to do their jobs independently. What seems to have happened here is a lapse in judgement while the reporter was sick and on a deadline. Pressure's a bitch.
2. This is certainly more of an opinion, but AI tools have zero place in any profession that relies on information integrity (law, journalism). The reporter shouldn't have used it, and the editors and editorial processes are not at fault, especially when Ars already forbids AI-generated content.
* I'm aware people think reporters are liars with agendas. You don't need to say it.
“Edwards also stressed that his colleague Kyle Orland, the site’s senior gaming editor who co-bylined the retracted story, had ‘no role in this error.’”
Has Orland issued a real apology? He bylined a piece containing fraudulent quotes.
Sad state of things. He did it because he was sick? That's close to claiming his dog ate the original quotes so he had to make some up.
Well, Ars Technica is already for quite some time on my ignore list, and this further solidifies its place there.
I clicked through the author's earlier stories when this first made waves. I obviously had no proof, but I was pretty certain that he's been using LLMs to generate stories for a good while.
When Ars released a statement saying this was an isolated incident, my reaction was "they probably didn't look too hard". I suspect they did, in the end?
"Don't believe everything you read on the Internet"
-Isaac Newton
The irony is Ars readers tend to be fairly anti LLM in the comments of every article
Man, the IRONY hurts. I love Ars for what they do and their comment section.
The readers there are borderline militant about AI's more problematic uses. This could have gone only one way.
Happy to see some accountability here. Athough it's unclear why the other co-author who stamped their name on that article was retained. Maybe they just stamped their name to meet their quota of articles. In any case this follow up action makes me take arstechnica standards a bit more seriously.
I bought a handmade arcade-style SNES controller from this guy one time. He should go back to making those, I guess.
> Edwards said that he was sick at the time, and “while working from bed with a fever and very little sleep,” he “unintentionally made a serious journalistic error” as he attempted to use an “experimental Claude Code-based AI tool”
I'm skeptical. I hate to be the one to say it, but I don't think this would have happened if he was using Claude 4.6 Opus.
This reads like “I was sick and my dog accidentally used AI to write my homework”
If the content is human written and you check your sources there is no way for AI to “accidentally” seep in. Sure you can use an AI tool to find links to places you should check and you can then go and verify sources. That’s obviously not what happened.
The headline is a bit sensational considering all we know from the reporting is that he isn't working there anymore. Fired likely, sure, but not for a fact.
I guess Blameless Postmortems haven't arrived in journalism yet.
Pretty weird that journalism as a business still revolves around "we hired a guy to write a thing, and he's perfect. oh wait, he's not perfect? it was all his fault. we've hired a new perfect guy, so everything's good now." My dudes... there are many ways you can vet information before publishing it. I get that the business is all about "being first", but that also seems to imply "being the first to be wrong".
I feel bad for the reporters. People seem to be piling onto them like they're supposed to be superhuman, but actually they're normal people under intense pressure. People fail, it's human. But when an organization fails, it's a failure of many people, not one.
Good time to watch Shattered Glass.
Imagine what he could have gotten up to with LLMs.
This is good. They had to distance themselves from a journalist who would do such a thing. But this is more or less on the editor I think. So let’s see if they learn from this.
I liked his articles about AI. They were generally quite good. He has an understanding of AI that usual journalists don't have. But to use an LLM for writing is deception.
I'm very bad with names and quotes, so sometimes I'll ask ChatGPT something like "what's that famous quote Brian Kernighan said about programming language names" and it will just make shit up, when really I was thinking about Donald Knuth. But according to ChatGPT, Kernighan famously said:
“Everyone knows that Perl is designed to make easy things easy, and hard things possible, but nobody knows why it’s called Perl.”
Which of course returns 0 results on Google.> ...he attempted to use an “experimental Claude Code-based AI tool” to help him “extract relevant verbatim source material.” He said the tool wasn’t being used to generate the article, but was instead designed to “help list structured references” to put in an outline. When the tool failed to work, said Edwards, he decided to try and use ChatGPT to help him understand why."
This, right here. Coming from an "AI Expert", this is what we can expect the future to be. One AI isn't working? Let's ask the other AI why. I have no words for that reflex. It's beyond idiotic. It takes everything that's human about your reasoning and tosses it aside. What a dumb idea.
> It also comes at a moment in which many media bosses are pushing staff to find uses for AI
Which should be a red flag in and of itself. You don't need to "push" people to "find uses" for genuinely useful tools.
If a tool is not fit for purpose then it either gets fixed or gets discarded/replaced.
AI is not a tool and from the way things are going never will be. Humans are more tool-like in that sense. In this case the human was discarded, the AI remains.
That was wise. It was an honest mistake, but a direct hit to is credibility that made not just him, but the paper, look sloppy. And in an era where people are deeply concerned about journalism pedigree.
I read the bluesky in article posted and Benj Edward's images that he had sent in bluesky.
The main comment I found relevant is probably this (There is more that he has written but I am pasting what I find relevant for my comment)
> I have been sick with Covid all week and missed Mon and Tues due to this, On friday, while working from bed with a fever and very little sleep. I unintentionally made a serious journalistic error in an article about Scott Shambaugh
... > I should have takena. sick day because in the course of that interaction, I inadvertently ended up with a paraphrased version of Shambaugh's words rather than his actual words.
> Being sick and rushing to finish, I failed to verify the quotes in my outline against the original blog source before including them in my draft
The journalistic system has failed us so much that in the news cycle, we want things NOW. I think ars-technica post went viral on HN as well before the whole controversy and none were wiser until Sam commented about the false quotes.
It prefers views and to get views you have to do work now. There is no room left for someone being sick and I think that this sort of expands to every job at times.
And instead of AI being a productive tool, it can act as a noise generator. It writes enough noise that looks like signal and Tada, none are the wiser.
People think that using AI with an person is gonna make their work 10x more but what's gonna happen is the noise is raised 10x more and the work of finding signal from that noise is gonna increase 10x more (I am speaking about employment related projects, obv in personal projects this might not matter if it might have 10x noise or 100x noise if it can just do the thing you want it to do)
When AI systems are constrained, they can deny you your api request with marginal loss. But when Human people are constrained, they really can't deny your employee's request without taking massive losses at times (whole day leaves) and I have heard in some countries, sick days can be a joke. This could very well be cultural because sick days are well implemented in Europe compared to america (from what I hear)
I don't know about Benj but some reporters are really paid peanuts. Remember the pakistani newspaper which had pasted Chatgpt Verbatim with content like "“If you want, I can also create an even snappier ‘front-page style’ version with punchy one-line stats and a bold, infographic-ready layout perfect for maximum reader impact. Do you want me to do that next?." WITHIN the newspaper.
I believe that humans should be treated with more dignity so that they feel comfortable around taking sick leaves when they are sick... or just fixing this culture that we have of people chugging along in sick leaves.
Until then, AI is bound to be used, I don't think that this is gonna be a single incident, and AI will produce noise/spew random stuff. Imagine you are a journalist and you are sick and you feel like there's a magical tool which can do the job for you when you are sick. You use it and in the moments of sickness, you are in the IDGAF attitude and push the article to main.
I personally don't believe that this is gonna be a single incident with this whole story playing out like this at the very least.
If any Journalist is reading this, please take sick leaves when you are sick. Readers appreciate your writing and I hope you don't integrate AI tools into your workflow (a lot) that the work is started being done by AI in this case. Even without AI I feel like you guys might not be working at the best mental space and Readers are happy to wait if you add unique perspectives into the story, something I don't think is possible when you are sick. If any employer try to still pressure you, just share this message to them haha to tell your employer what the people want (and what brings them money long term).
I also hate how the culture has become of finding the article which came the fastest after an event happens because that would promote AI use more often than not and it to me feels like jackals coming out of nowhere to try to take whatever piece you can take out of a particular news and that to me doesn't feel soo great of look. (I know nothing about how such journalism works so sorry if I am wrong about anything, I usually am but these are just my opinions on the whole thing)
Really disappointing. A lot of us have always considered Ars Technica to be the last of a dying breed of ultra serious, no-nonsense professionalism.
Obviously, we were rocked by the DrPizza scandal years ago...and now this.
Sobering.
people have said enough about the ethics of all of it but what I found even sadder is that the story made me curious to take a look at the actual piece he "investigated" with AI, it's this one (https://theshamblog.com/an-ai-agent-published-a-hit-piece-on...) This is btw a bit more than 1k words, which takes the average American reader, not senior journalist, ~5 minutes.
This whole story involved asking Claude to mine this text for quotes, which refused because it included harassment related content, then asking ChatGPT to explain that, and so on.
That entire ordeal probably generated more text from the chatbots than just reading the few paragraphs of the blogpost. That's why I think the "I'm sick" angle doesn't matter much. This is the same brainrot as people who go "grok what does this mean" under every twitter post. It's like a schoolchild who cheats and expends more energy cheating than just learning what they're supposed to.
"He added that, upon further review, the error appeared to be an “isolated incident.”
Yeah, no.
So this is another way how you can lose your job because of AI.
To me this is extremely relevant regardless of the specifics in this case. From TFA:
> It also comes at a moment in which many media bosses are pushing staff to find uses for AI — as are executives across most industries — even while clear guidelines around use of the technology that uphold editorial ethics remain elusive.
Anyone who's working with computers now knows this is true. We're being pushed relentlessly to use AI; in some cases (I've heard second hand) people are mandated to use AI, sometimes forbidden from crafting code manually, and are disciplined if they don't. Yet the guidelines are very unclear, as they must be since if we're honest we're all threading new ground.
This, being mandated to use AI at all costs yet given very brittle/unclear indications on how to use it, and these guidelines evolve weekly, and also we're all fearful of losing our jobs, makes for a recipe for disaster.
So yeah, this journalist should have called in sick and use better judgment when toying with AI tools, but still there's a wider problem and the responsibility for this craziness is also on the leadership of most companies and the investors pusing for this.
(None of this is an excuse for generating AI slop. I hate it and I don't need to be told any guidelines about not doing it. If you cannot be bothered to write the text, I cannot be bothered to read it.)
So the original blogger got slandered by an LLM agent, then got slandered again by a human journalist who used an LLM agent to write the article about him getting slandered by an LLM agent? How ironic.
But, does that mean he got slandered twice by an LLM agent or once by an agent and once by a human? Or was he technically slandered 3 times? Twice by agents and a third time by the journalist? New questions for the new agentic society.
> senior AI reporter
A true "senior" AI reporter should be more skeptical of LLM output than anyone else.
[flagged]
[dead]
[flagged]
Journalists and bloggers usually write about others’ mess ups and apologies, dissecting which apologies are authentic and which apologies are non-apologies.
In this incident, Aurich Lawson of Ars Technica deleted the original article (which had LLM hallucinated quotes) instead of updating it with the error. He then published a vague non-apology, just like large companies and politicians usually do. And now we learn that this reporter was fired and yet Ars Technica doesn’t publish a snippet of an article about it.
There’s something to be said about the value of owning up to issues and being forthright with actions and consequences. In this age of indignation and fear of being perceived as weak or vulnerable due to honesty, I would’ve thought that Ars would be or could’ve been a beacon for how things should be talked about.
It’s sad to see Ars Technica at this level.