So they’re not using it now, because it’s insufficiently trustworthy…AI isn’t going away, and I could imagine using it in the future if it becomes more trustworthy and perhaps if the companies pushing it find more ethical business models.
The editorial & review process is still my biggest curiosity about Ars in general.* Hardly the only writer who propagandises in their writing: others have called out Eric Berger & his very very pro-SpaceX writing, and I agree. I think its a challenge when a highly capable and connected reporter is also (presumably) an editor who can publish without oversight.
https://en.wikipedia.org/wiki/Singular_theyYou got “they” from one persons quote? Didn’t bother to read the article or spend two minutes to find out who the CJR is.
Take you one line hot takes back to X.
At this hour it is unreasonable to expect peak wit.Lol the comebacks are weakening in the thread…
I generally agree with your assessment, I would just reiterate my recommendation that ars should indeed properly research, verify and then publish a postmortem analysis capping off this incident, if possible also a related but separate article illustrating the greater issue as it concerns publishing in general now (since this would be a very different type of article it should remain separate but linked).Ars Technica does not fall under COPE guidelines because COPE is designed for journals that can afford to let a retraction take years. Ars chose to act in an afternoon. The editors/authors might have been on leave, spending time with family etc. Ars Technica is not a "breaking news" platform with 24/7 coverage. Given the choice, I'd rather have a clumsy, immediate "unpublishing" that stops the lie cold than a "zombie article" that misleads people for a generation.
Whataboutism completely unrelated to the breach of ethics at hand.Maybe this is karma for selling out our comments to OpenAI then disabling comment deletion to take away our only tool to avoid it.
Our words helped train the AI that produced the slop in the retracted article. Benji's not the only one with blood on his hands; thanks to Ars we all do.
I had respect for your position right until you said you were going to read the articles you no longer trusted anyway.That's it, I'm done then, I'm cancelling my multi year subscription. It was fun guys, but if your staff are going to use LLMs to write your articles without disclosing it (or even at all), I'm out. I'll go back to being a freebie reader with Adblock enabled.
I've found AI to be terrible as a search engine too, and a leading cause of the degradation of Google's search results in recent years.This has been the case with "AI" for a long time.
I can not count the number of instances where "AI" gave me very, very confident instructions on how to configure certain piece of software, or how to generate the configuration files or which options to use.
The only problem was - instructions were completely, and I mean completely wrong.
Not to mention that it would try convincing me that software used SQLite DB for configuration, while it was nothing like it. And so on, and so on.
I use "AI" as a very, very good search engine. It excels in this task.
Anything else - yeah, no. Not yet, at least.
I would just say the response is incomplete at this point, but the completion is just not possible before the legal fallout has been processed and before the evidence is fully analyzed.This really deserves a “I used AI uncritically and this is how I screwed up” follow-up article to accompany all of the ones about lawyers messing up their careers with AI citations.
Add me to the list of people who feel that the retraction was not handled at all well and the original article should have been updated.
For mine, google's full-force enshittification dates to 2019 and its strategy change to being an advertising company, no longer a search company.I've found AI to be terrible as a search engine too, and a leading cause of the degradation of Google's search results in recent years.
Hypothetically, you could use an LLM to create a software tool that downloads the content of a webpage, identifies the quote marks, and copies the text between them. That would not be subject to hallucination. You could also create a tool that asks the LLM to identify the quotes, then use a text search to verify a match.If someone is promising “only verbatim text and no hallucination” from an LLM, they’re lying. As this retraction illustrates.
If they were telling the truth, they’d be too busy swimming in their pool of VC money for their new legal analysis LLM (or alternatively, their big pile of equity from their LLM company of choice) to roll out random tools on the internet for free.
Leaving aside whether it’s doable at all, getting 99.9% reliable LLM output is the holy grail of the field and anyone doing it would be screaming it from the rooftops.
I don't mean to argue with your post, but to use it as a good clear statement to respond to: I think that constitutes malice, or something very near to it. If the Ars policy forbids using AI tools for an article, why were AI tools being run on article[-adjacent] text? What was he "testing", if the tool could never be used for his work product by policy already?Maybe you can rationalize it as getting confused about which code output which quotes and what guarantees there were supposed to be.
It's definitely FUBAR; but to me, there's plenty of reason to believe it was not intentionally malicious.
The only thing that's hypothetical about that is using an LLM to do it. I could write this in either JavaScript or PHP using RegEx to match patterns with a very high degree of accuracy and my degrees are in Communications, not Computer Science.Hypothetically, you could use an LLM to create a software tool that downloads the content of a webpage, identifies the quote marks, and copies the text between them. That would not be subject to hallucination.
The whole AI situation is so strange.
So the issue is really that Ars used AI hallucinated quotes but the whole AI agent attack and writing a defamatory article actually happened. At first when reading the retracted notice and replaced text in the original ars article, I understood it to mean that the WHOLE thing never happened.
My subscription hinges a lot more on ars' response to the writer's malfeasance!This is not enough… don't you realise that your reputation is at stake? Are you confident that your subscribers will not cancel, following this "incident"?
There's often a distinction made between using AI to analyze, understand and check content, as opposed to using it to generate content.I don't mean to argue with your post, but to use it as a good clear statement to respond to: I think that constitutes malice, or something very near to it. If the Ars policy forbids using AI tools for an article, why were AI tools being run on article[-adjacent] text? What was he "testing", if the tool could never be used for his work product by policy already?
Even the "innocent" explanation to me seems on its face to be an admission of AI tool use against Ars policy. If it wasn't, then I think the policy needs to be: No AI use from start to end, at any point in the process, period.
That this has happened here shows how insidious the temptation to use AI as a shortcut is, like root beer, and the Federation.
eh, that's not exactly what they said. they just said they don't want to financially support Ars if it uses LLMs, which could be for any number of reasons, like a moral objection to LLMs.I had respect for your position right until you said you were going to read the articles you no longer trusted anyway.
we can all agree, hopefully, that whatever happens to the author(s) involved as a result of this, it's a good thing they will have the union on their side to ensure they're properly represented and aren't being scapegoated to hide any issues that might exist in the broader editorial process.I think my biggest takeaway is…I think Ars’ writers are unionized and I think that’s good. My personal vindictive streak is very much in the camp of this being a fireable offense.
This is really where we're at. Humans associate polished communication with intelligence. A system that can produce text that sounds sophisticated and reads well is deemed "intelligent".We've known the explanation for years, now: we so strongly associate natural language manipulation with real human actors, that the illusion is very very powerful, and immediately bamboozles you the moment you let your guard down. Especially because it doesn't give away any clues, and has no internal pressure of any kind (fear, self-preservation, reputation...) to "not fail ypu".
Thank you for this post.At first I agreed with you, upvoted, and moved on. But then I thought about the implications of doing this in today's world.
5 years ago that would have been an ideal response. Because a human encountering a page of strikethrough text braketed with warnings could reliably be expected to interpret the article as intended.
However ai cannot be trusted to do this and therefore any misinformation contained in the article could be spread to unknowing humans by ai.
There needs to be a more careful solution.
They had enough time to blackhole the original article, replace (by recycling the URL) the content wholesale with a retraction notice but left the original text out as well as removing all the original comments. They disabled the comments section on theThe article was redirected to /dev/null within about two hours of publication on a Friday afternoon. We are still only about 50 hours out from that event. There have been basically zero conventional working hours since the failstorm erupted.
I am not going to say that the Ars editorial staff has necessarily covered itself in glory here--you could make an argument that this should have been an "all hands on deck, 6a-6p work, Christmas is cancelled" event--but to me it does not currently seem to be dripping with concentrated asscoverium.