I feel what Benj did was much then what you're leading on. The mis-quote was from a website, not from an interview. There is absolutely no reason why Benj could not copy the quote from the AI system, paste it into a "find" box on his web browser, and confirm the quote exists. That takes 30 seconds to do. Benj did not do that. Benj has 20 years of journalism experience (per his website) and I am stunned he failed to perform such a simple task.It is a fundamental part of a journalist’s responsibility to verify their quotations. I believe that the journalist should often go further and ask if the quoted person has a further comment.
There's an old saying in Tennessee--I know it's in Texas, probably in Tennessee--that says, fool me once, shame on--shame on you. Fool me--you can't get fooled again.well now that they got caught they'll make extra-certain they don't get caught again
The only criticism I have for Ars itself is pulling the article and comments down. I shouldn’t have to go sleuthing to find the original article as it was published.
Expecting integrity and transparency is reasonable. Expecting perfection is not.I've enjoyed Kyle Orland's work over the years but I won't trust it going forward and honestly will be moving on from Ars as well. This is not the first issue with intergrity they've had and at some point as a reader you have to admit there is something broken with the culture regardless of how much you enjoy the content.
Decision making requires judgment. It's a silly hair to split.Errors in judgement should be forgiven. Errors in morals should not be forgiven. I do not believe that Benj made an error in judgement, but instead he made an error in morals.
1. He (SHOULD!) know, as an AI expert, that AI systems are probability generators, and can not think or make decisions. Because it's a probability generator, by definition, it can not extract quotes from websites. If there is such a tool created, that's worthy of an article itself.
2. He is a journalist of over 20 years. The first thing journalists are taught is to always verify the quotes. The quotes, in this case, come from a website, which could easily be verified using the "Find" command on any web browser. He failed to engage in that very simple task.
3. Because he used AI generated quotes, he put his employer at extreme legal risk. Thankfully the quote generated by AI wasn't harmful, and the person wronged took it in stride. However, there is no reason at all why he didn't verify the quote. But could it happen again? Would an employer want to take that risk?
Clearly, he is not a new journalist, and he is not new to AI. Either he's incompetent as an AI expert, and if so, dismissed from the AI beat - or incompetent as a journalist, and in that case he should be dismissed from working for Ars.
Why do you believe this is an error in judgement?
My take on Clawbots is simple. It's humans pulling the strings.But the repercussions of clawbots mean that eventually someone will have to take legal responsibility for their actions.
Agreed. The soccer rule of two yellow cards (warnings) then a red one (out!) ought to apply.
[...]
Well there are also the other villagers saying "I'm not sure forging quotes in a journalistic piece is such a big deal".
This is an interesting take - the quotes were made up bullshit, but were "truthy enough" that it doesn't matter to you that they were fabricated.
Please, someone here explain how at least 21 Ars readers can think it is okay for a journalist to misrepresent a key quotation. To my thinking, once is enough for a true journalist to post that kind of lie.
fabricate | ˈfabrəˌkāt | verb [with object] 1 invent (something) in order to deceive
forge | fôrj | verb [with object] 3 produce a copy or imitation of (a document, signature, banknote, or work or art) for the purpose of deception
lie 2 | lī | noun an intentionally false statement
I think "reckless disregard for the truth" is the applicable standard here. Putative AI expert asks the slopbox to pull quotes from a single blog post, an act which would take that expert maybe five minutes of human time, and he just rolls with it and publishes it to the world.I don't believe it's likely that Mr. Edwards intended to quote the subject of his article incorrectly, therefore I don't think those adjectives accurately convey the situation.
So back to intent - does it matter? My initial reaction is that it should matter, but I'm open to other thoughts on the matter.
I think it depends on how new a journalist is to the job. For a new journalist, I believe intent should matter, and a new journalist should be taught best practices - which includes not using AI tools to extract information from websites and then quoting what AI generates because these tools are just a word probability generator, and to reinforce that sources need to be tripled checked to minimize liability.So back to intent - does it matter? My initial reaction is that it should matter, but I'm open to other thoughts on the matter.
And a note to our current administration in DC - this is what transparency looks like.
It's not like Benj is shy of his AI usage as a tool to write better articles.
View: https://youtu.be/1nEph7-Viyc?t=255
In his interview with Ed Zitron, he is very candid, and he admits that GenAI is a good tool to help him fight against COVID Brain Fog. This is not a new excuse from him, the interview is from 4 months ago.
There's another definition of "fabricate" which is entirely devoid of intent. And he used a piece of software which fabricated quotes in that sense of the verb.Does intent matter in this situation? There are lots of accusations of "forging," "fabricating," "lying," and so on throughout these comments, and every single one of those requires intent to deceive. To whit, from the macOS dictionary:
I don't believe it's likely that Mr. Edwards intended to quote the subject of his article incorrectly, therefore I don't think those adjectives accurately convey the situation.
So back to intent - does it matter? My initial reaction is that it should matter, but I'm open to other thoughts on the matter.
Take the governor off a self-driving lawnmower, unleash it in Aunt Martha's garden, and watch the fun!My take on Clawbots is simple. It's humans pulling the strings.
We have reviewed recent work and have not identified additional issues. At this time, this appears to be an isolated incident.
While Edwards may not deserve it (I agree with you that he probably doesn't deserve it) ars employees, readers, subscribers, and sources deserve a publication and team they can trust. It's hard to have that trust when you keep someone on your team that violated it.I don't care about Jim Salter. He's not the one involved here. Edwards is and he doesn't deserve to get fired for it.
I absolutely agree with this but that is, again, a management problem. They took the article down instead of having the guts to keep it up with corrections and an explanation.While Edwards may not deserve it (I agree with you that he probably doesn't deserve it) ars employees, readers, subscribers, and sources deserve a publication and team they can trust. It's hard to have that trust when you keep someone on your team that violated it.
Yeah, what does Jim Salter know about what it's like to be a writer for Ars Technica, anyway?
Just to make sure we've left no one out, Thad Boyd was referring to the fact Jim Salter has written for Ars Technica and elsewhere (good stuff, too).I don't care about Jim Salter. He's not the one involved here. Edwards is and he doesn't deserve to get fired for it.
A concern I have is this. The article was written by Benj, with Kyle being a co-writer. Why didn't Benj tell Kyle that he was under the weather and just do a quick check on his sources and make sure everything is fine because he was using a new AI tool to extract quotes?So, someone who should and does know better is, simultaneously, trying to stay on top of things and gainfully employed in a rapidly-evolving field despite struggling with long-term health (including mental health & focus ability) drawbacks from COVID.
This is kinda where I've ended up. There are two wrongs here that got blended into one.It is a fundamental part of a journalist’s responsibility to verify their quotations. [...]
That Edward’s may have violated an Ars policy concerning AI is a distraction.
Not if it gets... deleted!In a non-related future story, this retraction seems to be on track for the yearly Ars "Most commented" articles for 2026 they post in December. Yikes.
I understand your strategery there...There's an old saying in Tennessee--I know it's in Texas, probably in Tennessee--that says, fool me once, shame on--shame on you. Fool me--you can't get fooled again.
Agree 100%. I was going to unsubscribe the last time they pulled this shit and turned off auto renew, but it renewed anyway. Maybe I screwed it up so let it go at the time.It's not plausible to me. This is not their first rodeo. This isn't even unprecedented in the last 5 years.
I've been loathe to dig up examples because I don't want to be accused of rehashing old points of contention and some of the topics are really sore spots, but this isn't the first article they've nuked (comment thread and all) and their non-response burned a bunch of goodwill and caused a small exodus of readers, subscribers, and community regulars. Ken and Aurich have had to step in then, too. I was one of those calling for a public accounting of the process in the aftermath and certain changes to the way Ars handles things as a matter of course, and we never got it. The changes that were made (which likely directly resulted in Aurich being able to lock things down and get an emergency editorial action so quickly this time) didn't extend to public disclosure about what went wrong and what measures were going to be taken to prevent the issue from happening again.
The "vibrant" discussion you're reading now might even be tame in comparison to what went on just a few years ago when that shit went down. But for many in the community, including dozens I'm still in touch with after they left, the wound was left to fester. Compounding the original mistakes was a sense that the incident was being papered over in hopes that it would just die quietly in the neglect and everyone would just move past it. Well, a lot of people did: they move right off of Ars because of how it was handled.
So I'm not going to cut slack if the whole thing ends here. I've been there before and it's awful. Worse is the implication that after last time, management didn't change its approach. It would be tragic if the same thing happened again when it doesn't have to. That would tell me everything the leadership of Ars really thinks about its readership, let alone the active community. That's why I'm hoping the powers that be at Ars have learned from before and will make different choices that are more responsive to the needs of their readers this time; I have to say that some recent smaller dustups have not left me 100% confident. There's a lot on the line as I see it. At least, there's a lot on the line for me.
This. If a YouTube video contains a bunch of DALL-E graphics, it's slop even if an actual human is in the video doing narration or something.
This became an issue with dubbed animation too. People didn't like the AI-generated audio. The audio was slop even if the video elements attached to it weren't. Therefore the product was slop.
Something like this is sort-of tenable with the "indie web" of small personal websites that still exist, e.g. Neocities.
Sadly the human affiliate content mill crap made it almost impossible on the Internet at large even before press-button slop generators were available.
These days I'm finding myself doubting general information pages that show up in specific searches unless they predate 2023 or thereabouts. Everything after that I treat as "probably slop." Like, if I see a cool plant growing by the road and want to know if I can find a spot for it on my property? A search might turn up hundreds of pages from the last couple of years on exactly that topic, even that specific plant... which itself seems suspicious.
The post prior to yours called it "reckless disregard"; without trying to get into legal standards of negligence I'd lean towards a casual disregard for basic journalism principles. The guy seemed to be responsive to interaction on this--and not incidentally--said in a follow up note that Ars was not one of the sites who had reached out to him. The heck with a web search. Why not send him a note and request for comment?I think it depends on how new a journalist is to the job. For a new journalist, I believe intent should matter, and a new journalist should be taught best practices - which includes not using AI tools to extract information from websites and then quoting what AI generates because these tools are just a word probability generator, and to reinforce that sources need to be tripled checked to minimize liability.
Benj, according to his website, has over 20 years of experience in Journalism as a reporter, and is an expert in AI. This tells me the following:
1. He failed to performed basic due diligence in confirming a quote from a website is valid (which could be performed using a copy/find/paste in the web browser of his choice).
2. He does not understand that GenAI tools are a probability generator, and can not accurately extract text from websites (a massive failing to understand how GenAI operates!)
As I understand it, a reporter should always verify the source three times, just to make sure the source isn't being misquoted. This minimizes the risk of liability.
If he can't do a 30 second search of a website, that shows a massive failure in willingness to verify AI information.
Lots of thoughts...
Were I to guess, the "process" in this case probably has the writer as a single point of failure. I would not call that ideal, but (1) I'd be surprised if it was particularly unusual, especially in the current media environment, and (2) even places with "good process" probably run into this at times. For the second point, take a journalist quoting a confidential informant. Unless there is absolutely always someone else there/on the phone (something a confidential informant would likely want to avoid), you have to take the journalist word on any quotes they collect.
I would encourage a more in-depth follow-up once folks are back from holiday.
Probably the first thing it would be nice to hear is an overview/rational of how you make the call to take a problem piece down versus leaving it up. Honestly, I can imagine arguments for taking either approach in a given case. I would offer that I don't think you have to convince everyone of the right answer in any given case, but I think some folks might feel a bit better if they heard some rationale that seems solid.
From the retraction:
These two sentences do a lot of heavy lifting, and I do not mean that in a facetious manner. Here's my mental model of how that worked:
- Mobilize a team to work over a holiday weekend.
- Select a healthy sample of articles from the writer, with a heavy weight on the weekend ones.
- For each article, compile a list of all facts and quotes.
- Verify each fact. For quotes from public sources, verify those.
- For interview-based quotes, contact the quoted parties to assess the veracity of the quotes in the article.
- If you complete this for a healthy sample of your healthy sample and do not find issues, you can probably conclude that the issue was not pervasive.
If you make it here, you can probably reasonably conclude that there is not a pervasive problem in the writer's work, and dismiss folks for the rest of the holiday weekend--to come back to the rest of the writer's corpus when the holiday weekend (and ideally, some comp time) is over.
There's every possibility I'm wrong on some of the, "what," here. However, it went, I think a lot of folks would like to know how something like this gets handled. That's arguably of extra importance if there is a single point of failure vulnerability here (which seems likely) as well.
I think the, "why," matters, too. It makes sense to set the first priority in the crisis response would be triaging how much bad material is making it into articles and start pulling down problems if they are found in any number. If a good sample does not suggest a serious problem, though, I can see letting people have their weekend and taking more time to thoroughly vet things during business hours.
As of this time, the writer does not seem to have been dismissed to this point. That feels right at this time. I think it's prudent to review the situation carefully, and I think it's appropriate to have that happen primarily during work hours (since there does not appear to to need to be a flood of retractions), as well as recovery if illness is involved. I think also think it is prudent to get some, "trust but verify" information from the writer involved, and it is reasonable that said writer be paid for the time that takes. You can't readily do that if you dismiss them hastily.
In terms of consequences, I think the first thing I would like to hear is that any work involving that writer is on hold for the duration of the investigation. (This is a good time to say what shouldn't need to be said.)
At the conclusion of the investigation...it's tricky. On one hand, I'm not out for metaphorical blood, but on the other I do not have a quick thought on what, short of a dismissal, comes down firmly enough to reinforce editorial standards. (Editorial standards aside, I'm sure a lot of other folks already went through the wringer of the weekend digging for more potential problems, and more still will do the same when they get back, many of them through no fault of their own. It's tough to go light when there is so much impact on other employees.) I suspect I'm not alone in those feelings. If you come up with a response that is short of dismissal and explain it, I would hear you out and suspect others would, as well.
PS: This is at least the second time I've seen Aurich jump on something like this, and as rocky as they are, I suspect they are far better for his jumping in. He's one of the great folks you have, and one of the keenest disappointments in such messes is how it sidelines such great work.
My employee did bad thing X, when I found out and confronted them he asked me to do thing Y, I did thing Y. I am not fully responsible for doing thing Y because my employee asked me nicely to do it.One thing that most everyone is seeming to miss or at least not take into account when complaining that Ars took down the article is that in Benj's BlueSky mea culpa, he said that he asked his boss to pull the piece because he was too sick to fix it.
Maybe that's revisionism on his part, but if he's saying he asked it to be pulled maybe not put all the blame on Ars for pulling it?