Even to a fairly new Ars visitor like myself, Benj Edwards has a clear reputation for AI-centric pieces that read like Press Releases not journalism (it gets widely commented on in every article's comments section). While I agree with others that it's important not to jump to a conclusion, I feel like one of many people primed to assume Benj's AI fetishism got out of control this time. Considering a visible history of bias in favor of the AI companies in his articles, if he is responsible for this I would hope Ars considers that in any disciplinary actions. I don't think this is a one-off incident.The original article is attributed to Benj Edwards and Kyle Orland. The fact that the article did not meet the journalistic standards of Ars and got retracted is important. But the stance of the authors is equally important because it is a reflection on their future contributions and whether we should trust them.
One might be forgiven for naively thinking that this would only get 96.667% positive feedback. Thing is, some of us would be delighted to lose a little blood while saying hello to a new bobcat friend...If you substitute bobcats for office chairs one time in every 30, you can still have 97% positive feedback.
The bobcat may not be delighted though.One might be forgiven for naively thinking that this would only get 96.667% positive feedback. Thing is, some of us would be delighted to lose a little blood while saying hello to a new bobcat friend...![]()
I think it's the most accurate summary of what AI is contributing to society yetHonestly, if someone told me this story I would have laughed and said it was too silly to be believable.
An AI bot submits a pull request that gets rejected, so that bot goes off the rails and publishes a blog ranting about how humans are prejudiced. The person covering the story is the lead AI editor, and uses an AI bot to help in the publication of the story, but the bot fails. The lead AI editor then turns to another AI bot to explain why the first bot failed, and gets caught using hallucinations in the published story.
It's just... too ridiculous.
a.k.a., "I only stole from the company once"if it is verified as an isolated occurrence about which everyone is honest within a short, but non-zero, amount of time
Someone who’s mind has the ability to change given new information.
While I get the point you're trying to make, comparing this to vehicular manslaughter (of children no less) is way off base.It's very sad to see Mr. Edwards seemingly throw away his career like this, but that would be the only reasonable and acceptable outcome. "Sorry I was driving tired when I drove my bus over a crosswalk full of schoolchildren". Well... yes, but passersby were amazed by the unusually large amounts of blood. Passersby were amazed by the unusually large amounts of blood. Passersby were amazed by the unusually large amounts of blood.
Sometimes you walk old battlefields or ancient European cities and see intentionally-unrepaired bullet holes in columns and monuments. "We left it like this so we wouldn't forget". I hope that's the long-term outcome for Ars Technica--made stronger by a moment of weakness.
Now we can see how "AI" is destroying jobs...!Honestly, if someone told me this story I would have laughed and said it was too silly to be believable.
An AI bot submits a pull request that gets rejected, so that bot goes off the rails and publishes a blog ranting about how humans are prejudiced. The person covering the story is the lead AI editor, and uses an AI bot to help in the publication of the story, but the bot fails. The lead AI editor then turns to another AI bot to explain why the first bot failed, and gets caught using hallucinations in the published story.
It's just... too ridiculous.
Hmm, your "obvious" does not match the BlueSky post where the author talks about the first time use of an AI assistance tool (which failed).What happens to the writer who has obviously been shovelling AI slop since well before he got caught?
i am also curious as to how it actually happened. not to shame writers, i'm just curious about it.Yes it was actually pointed out by the person who the quotes were attributed to. He showed up in the comments of the article, it quickly caused a furor. Removing the article was the right call.
Two different writers were listed on the byline. Did one use AI without the other's knowledge? Seems like a big lapse in judgement happened somewhere.
i think it's good they removed it, but i think it should be covered in more detail.It was caught because the author of the blog post, Scott Shambaugh, posted in the comments that the 2nd half of the article was made up, that the quotes attributed to him were fabricated. Other posters who had read his blog post, affirmed his statement was correct.
I'm glad ars acknowledges the error, but removing the article and shutting down all commentary strikes me as counter productive, as far as transparency goes. It's not like there haven't been other controversies before.
This wasn't a mistake. It was a lie- a deliberate attempt to deceive readers in direct violation of company policy. I make mistakes at work and I get a talking to, which is reasonable and fair. If I lied publicly, misrepresenting my company, I would be terminated, and I would expect nothing lesswould hate to think that either author would lose their job over this - we all stuff up at some point; it’s what we do after that point that defines us
It’s embarrassing, and really poor judgement. Summarising an article is one of the things these chatbots are supposed to be good at - funny that it can’t even do this. Both authors should already know that!
When Glass made up a fake company for his fake story, he produced a website, business cards and flyers for the company participating in the conference he made up. When caught, he had his brother pose as an executive of that company to try to fool his editor. If that level of dishonesty were going on, I would hope for that level of response.
This appears to have been a lazy and careless use of AI. While absolutely unacceptable and galling when writing for a site that has frequently cautioned that AI cannot be trusted, it is far from the level of dishonesty that Glass employed.
Likewise. It's been pretty full-on, and the vehemence is something of a surprise over the Monday morning coffee!Don't usually post, but posting now to express appreciation for both author's work and observe that I can think of lots of scenarios consistent with posted statements that would not make this close to a firing offense.
I know that there is a real sense of betrayal, given that (for me at least) Ars is generally a bastion of standards and sanity, but if it is verified as an isolated occurrence about which everyone is honest within a short, but non-zero, amount of time, I can't see it as a nefarious plot, and I'm a bit surprised by the instantaneous vehemence here.
The problem is that whatever "do not crawl" flags are put on the original article it is going to get "ingested" (e.g. stolen) by LLM crawlbots, at which point it will end up in the LLM database and will inevitably resurface at some point adding to the original LLM operator's false accusations against the victim. I don't see any alternative to taking it down, even knowing it has already been copied by the Wayback Machine.I'm not going to read all the comments right now (working), so this may have already have been covered, but there are a lot of comments about not leaving the article up with a retraction. Personally at this point in time I think the article should have been named, but not linked.
Be as transparent about the process and results as you can, when you can. I expect that there was an urgent meeting with the lawyers where it was stated "this is the limit of what you can say at this point".
Whilst I agree that having the article + retraction is the preferred situation from a human PoV, we are also living in a world where that incorrect information is then being indexed and sucked up into LLMs, making the problem worse. So initial damage control is IMO probably the right move - esp. on a holiday weekend. Same thing that should be done with a bad production release - revert and contain the damage as best as possible in the shortest time possible, then do your root cause analysis and mitigations to whatever came out of that.
A journalist fabricating quotes from a real person (or fabricating evidence more broadly) is nearly the worst non-malicious professional act I can imagine within their field, just as accidentally running over a bunch of children is nearly the worst non-malicious professional act a bus driver could perform. The implied equivalence was completely intentional and I stand by it confidently and resolutely.While I get the point you're trying to make, comparing this to vehicular manslaughter (of children no less) is way off base.
It was Benj, and according to him, COVID made him do it:
View: https://bsky.app/profile/benjedwards.com/post/3mewgow6ch22p
The author has damaged his credibility by publishing fabricated material. Any rationale for publishing those fabrications should be met with great skepticism in light of the exact fabrications he’s attempting to rationalize.Hmm, your "obvious" does not match the BlueSky post where the author talks about the first time use of an AI assistance tool (which failed)..
Man, the "I was sick" excuse. What's next? The dog ate his sources, so he had to use AI instead?
i am also curious as to how it actually happened. not to shame writers, i'm just curious about it.
I agree that your implication was intentional. I just think it's hyperbolic at best. The scale of potential harm of a journalist fabricating quotes from a blog comes nowhere near that of a bus driver causing a fatal accident. You're comparing a professional mistake that will at most will result in his firing to something that's a criminal offense.A journalist fabricating quotes from a real person (or fabricating evidence more broadly) is nearly the worst non-malicious professional act I can imagine within their field, just as accidentally running over a bunch of children is nearly the worst non-malicious professional act a bus driver could perform. The implied equivalence was completely intentional and I stand by it confidently and resolutely.
Not to the point of doing something that was clearly wrong like this!Have you honestly never in your life had the experience of making a poor judgement call and/or executing something poorly because you were sick?
I will continue to beat this dead horse- lies and fraud, in direct and deliberate violation of company policy, are not a "poor judgement call". Most companies would call that "gross misconduct"Have you honestly never in your life had the experience of making a poor judgement call and/or executing something poorly because you were sick?
It would be much sadder if 20 (?) Ars Technica journalists lost their jobs because nobody had a whisker of faith in the journalistic standards of the organization and the whole thing went under. Or the other writers at any other organization that foolishly hired him into a like role after this.But as a human being, I also hate the idea that a screw-up while sick, which was likely a quick "this will only take a minute and nobody will know," has thoroughly fucked Edwards's job and reputation. He should have known better - he's clearly spent a lot of time around these abominations, as his job, and should be well aware of their pitfalls. But I feel sorry for him, and yeah, I hope he has a chance at professional redemption, either here or elsewhere.