I saw it in my own Mastodon feed from the arstechnica flipboard account. Didn't see it in the mastodon.social feed. I'm honestly not sure if Ars even runs the "arstechnica" profile on flipboard, but it looks official, so maybe that's what you remember popping up in your feed?Hey, uh, I'm pretty sure I found the retraction through my Mastodon feed, and boosted it. But now that post is gone. Did you actually delete the mastodon post+thread for YOUR OWN RETRACTION?
This is feeling much less transparent than I wanted to believe. Thank the gods for RSS.
Yeah, this is reasonable and something I will spend some time thinking on. But actual, legit infection(or poison or whatever)-induced-literal-brain-malfunction, causing someone to do something insane in the literal sense, would be a genuinely ameliorating factor if the organization then acts in a systemic way to prevent that being a SPOF again. I can see that being a blind spot for an org that starts fairly young and in tech and it's something I've contemplated on as well as I age and watch my parents (and before they passed recently grandparents) age. We are all running on actual physical wetware at the end of the day, and while normally that's reliable enough we can act as full abstraction, the risk of malfunction and degradation is real, even for the best of us. As in the airline industry and so on it's up to the organization to help deal with the reality of messy humans in the loop. Maybe here there needs to be a much stronger policy around "when you're at all sick, all deadlines are suspended" or something.I'm willing to extend a tiny bit of grace here. Guy whose job it is to learn about and report on AI finds out about a new tool that is advertised as being suitable for extracting verbatim quotes and decides to try it out, in a haze of illness he fucks up very badly.
COVID often really fucks with your cognitive function, the first mistake was to continue working when sick, and that mistake can be partially attributed to the illness itself, as can all the following mistakes.
This explanation, even if you remove COVID from the equation, is slightly less troubling than what I had assumed.
I'm not sure what the best way to move forward is, personally I would like to continue reading reporting by Benj Edwards here on ars.
Not trying to be a jerk but this kind of overly-lenient attitude among those experimenting with LLMs needs to stop. The bolded parts are literally contradictory....I have started using Gemini quite a bit in the last 3-4 weeks, and it is shocking how good it is and how much detailed information it can give me about obscure topics like particular revisions of automotive parts or configuration settings.
What's also eye-opening is just how often it's completely wrong.
...and when it's incorrect, it's confidently incorrect.
[snip]
[and then] It's often still incorrect.
They were specifically requested not to comment until Ars management had said their piece in public. Presumably, once the retraction note was posted, they were clear to make their own statements public.I appreciate Benj's clarification, but -- aren't both Benj and Kyle now taking to Bluesky to comment on a situation on which they were specifically requested not to comment?
That just means the PTO is inadequate. I know that's very common in the US, but that doesn't mean it's right.I don't know what that policy is, but as someone who spent many years working remotely, you just don't want to waste your PTO on being sick. You probably are going to sit in front of the computer anyway because what else are you going to do, you aren't going to infect anyone, and usually you think it's better to just muddle through rather than spending a limited resource or let down your co-workers.
I don't know that this is necessarily always the right attitude, but I think it's very common.
This has been the case with "AI" for a long time.The whole situation is so strange. I have started using Gemini quite a bit in the last 3-4 weeks, and it is shocking how good it is and how much detailed information it can give me about obscure topics like particular revisions of automotive parts or configuration settings.
What's also eye-opening is just how often it's completely wrong.
...and when it's incorrect, it's confidently incorrect.
It has suggested parts that fulfill my requirements that simply don't exist, and it has explicitly told me which programming values to change to update the 12V battery configuration in my Mach-e, and when verifying, those values are in the wrong location.
It even explains why those are the correct parts or correct values.
When I correct it, it says, "well spotted! Those are the correct values for the F-150 and x other vehicle. The correct value for your vehicle is 'y'".
It's often still incorrect.
It's so important to verify these things.
Is this AI?This is great to hear -- positive reinforcement for handling a mistake promptly and responsibly. Thank you for rewarding good behavior!
It will be interesting to see what follow up comes from the editors. I agree with others that this statement is a good start, but I want to see more. Why did this happen? Were other articles affected? Perhaps even a word from the people involved, explaining why they did this.
1. That's not how an article "retraction" works for most credible media outlets; that's how you cover a mistake by deleting it (trying to, thank you Internet Archive) and making some PR statement as a band aid on a bullet wound1. It seems they took the article down quickly, they just took time to post this article.
2. Ars probably doesn't know what will happen since it's a holiday weekend, and they probably haven't even had a chance to talk with the authors. Dealing with things like this take time.
3. The article literally says "At this time, this appears to be an isolated incident." Which we can infer to mean they believe only 1 article is impacted.
In other words, you're going to ragequit because of what you read, and then keep on reading.That's it, I'm done then, I'm cancelling my multi year subscription. It was fun guys, but if your staff are going to use LLMs to write your articles without disclosing it (or even at all), I'm out. I'll go back to being a freebie reader with Adblock enabled.
Indeed, the old saying "it's not the crime, it's the coverup" has stuck around for good reason. In this case, putting the notice of retraction, link to retraction article(s), aI think I'm even more disappointed in the way the retraction was handled than in the original error.
I’m more open to AI than many Ars commenters, I think, and using an LLM workflow to pull quotes and messing it up is a human mistake I can imagine making. That wouldn’t justify it (especially not checking the accuracy yourself), but it’s imaginable. I may be a bit biased as a fan of Benj’s video game history work long before he started writing for Ars.
But the weasel retraction is really poor form. Not linking the original article, obscuring the original article, hiding the names of the authors, and deleting the comment thread? Transparency this ain’t.
Errors happen and employees screw up. But hiding those mistakes, especially in journalism, is really bad for trust.
{strike} around the whole original article and delisting it, but leaving it up with the comments thread up as well (but locked from further replies) seems like a better response. Nothing to hide, no links broken, people can see what happened, while also having it be very clear that it's retracted.Maybe with the ever rising tide of AI, journalism organizations will start having to treat people more like aircraft pilots or the like, where if someone is sick enough it's effectively a "safety risk" and they should just be outright forbidden from any further work (enforce it technically too! disable the VPN etc) until they're recovered. Or at least any public facing work, maybe doing some equivalent of desktop clean up is ok, but nothing in the hot seat. Both for them and for the org. It's true that a lot of people might just try to work remotely normally but perhaps enforcing stronger work/life separation (you will take some time and relax and you will like itThat just means the PTO is inadequate. I know that's very common in the US, but that doesn't mean it's right.
Really, sick leave should be unlimited and coordinated with the company disability plans (so disability can take over if the employee is sick enough to be out for an extended time). It's absurd that we accept having a limited number of days to be sick. I'm fine with having reasonable guardrails to prevent abuse (requiring a doctor's note for extended absences).
I wouldn't attribute that to Ars. It's extremely common in white-collar work to just soldier through sickness, and that's only been made worse by remote work (ironic, given its origins in COVID). I know I've worked from home plenty of days when I probably should've just taken advantage of a sick day.This is also concerning insomuch as it suggests Ars staff may not have adequate sick leave, or may have a culture that discourages its use.
The policy already in place is that no AI stuff gets used without full disclosure, which was violated. So they can 'promise' again, but it's really all up to the authors to say they won't use AI and actually comply with that. The editors can't fact check every single thing in an article - if they could, they wouldn't need authors, could just write it themselves. So a reasonable sounding quote from someone goes right by, because why would an author just make that shit up when they're supposedly not using chatbots?Alright, thank you Benj for acknowledging this. Personally, what I'd like going forward is a promise that such "tools" won't be used again, and I don't think that's an unreasonable request.
There is an almost zero chance of a lawsuit. While false, the quotes did not in any way harm Shambaugh's reputation. Plus it would be incredibly hard they were made intentionally rather than through stupid choices. The article overall was very supportive of Shambaugh.2. "Probably"? Unless they have some really unprofessional people working for them they damn well better know as a lawsuit from Mr. Shambaugh I'm sure was on the table, if not still is.
The only way to avoid this is to just not use AI, which is pretty easy; we all managed to do it before a couple of years ago. Many of us still manage to do it today.Reading Benj’s explanation, I was going to say something to the effect that the only way to avoid this would be to have two computers, one where you can use AI for research, and another for the article, and you can never cut and paste between the two… but even that is not enough. You could just look at the fabricated quote and type it out with your hands.
It’s only going to get worse as AI is inserted everywhere. It’s in freaking notepad.
That's my prerogative yes.In other words, you're going to ragequit because of what you read, and then keep on reading.
Did you even read the statement? It literally says "We have reviewed recent work and have not identified additional issues. At this time, this appears to be an isolated incident."Has the author's previous work been reviewed? If there were no editorial safeguards in place to catch this, and it took the misquoted person coming to the comments, is there any way to know if it's an isolated incident?
We only know that Kyle was asked not to comment.They were specifically requested not to comment until Ars management had said their piece in public. Presumably, once the retraction note was posted, they were clear to make their own statements public.
I don't think it's unreasonable for Benj to make a public post taking responsibility for the screwup, especially if it's after management has said, "right, we've said our piece, go nuts". It's the responsible thing to do, especially since there are two potential culprits.
Now it's for Ars management to look at what Benj (in particular) has to say - both in public and in private - and decide what to do, and how much to make public about the process. My view is that more transparency is better, with the caveat that there are good reasons to keep some of the details behind closed doors, sometimes.
I mean with other authors, not just this one. Failure of the editorial process.Did you even read the statement? It literally says "We have reviewed recent work and have not identified additional issues. At this time, this appears to be an isolated incident."
The author lied to his editor and to us. He broke the rules of his workplace and damaged the public image and credibility of his employee in the process. I hope he reads these comments and recognizes how serious this is. I hope he sees that many of us are considering unsubscribing because of his dishonesty. I hope the management also read these comments and use them in the sit-down chat with Benji as evidence of why his employment has been ended- costing the company money through fraud and dishonesty. These are grounds for termination at any companyConsidering that the authors reads the comments it's disappointing that they haven't taken the message conveyed to heart.
Right, Ars are not obligated to our subscriber money. I'm not going to quit reading, but I've also set my subscription to non-renew.That's my prerogative yes.
A policy like this incentivizes employees to work sick, which means poor quality work is incentivized. Poor quality work should not be incentivized.I am starting to be less pissed off at Ben for putting fabricated quotes into his article and more pissed off at Ars for making him choose between working while having COVID and having less PTO for recreation. (I know that Ars is not the only company that does this, but it is a stupid policy that results in incidents like this.)