Being too nice online is a dead giveaway for AI bots, study suggests

msawzall

Ars Tribunus Angusticlavius
7,395
Instead of relying on subjective human judgment about whether text sounds authentic, the framework uses automated classifiers and linguistic analysis to identify specific features that distinguish machine-generated from human-authored content.
We asked a computer if another computer sounded authentic, and the answer was a resounding yes!
 
Upvote
142 (146 / -4)
Post content hidden for low score. Show…

chanman819

Ars Tribunus Angusticlavius
6,711
Subscriptor
When prompted to generate replies to real social media posts from actual users, the AI models struggled to match the level of casual negativity and spontaneous emotional expression common in human social media posts, with toxicity scores consistently lower than authentic human replies across all three platforms.
Time for Microsoft to dust off that old Tay code.
 
Upvote
170 (172 / -2)
…actual humans on social media keep proving that authenticity often means being messy, contradictory, and occasionally unpleasant.
What about spouting absolutely bullshit with perfect confidence, never using one word when an inaccuracy followed by a dozen rationalizations will do? The great similarity between AI driven chat and a human being is the almost complete inability of either to tell the truth. The most significant characteristic of chat AI is its perfect resemblance to a dirty politician.
 
Upvote
47 (52 / -5)
Post content hidden for low score. Show…

UserIDAlreadyInUse

Ars Tribunus Angusticlavius
7,719
Subscriptor
"Did you run that through the AI checker?"
"Didn't have to. It's human."
"What? Look at it. Never uses the same synonym twice, em-dashes all over the place....what makes you think it's human written?"
"In the essay I was told where to go and how to get there. Twice. And conjectured about certain proclivities that my matrilineal family line engage in with paying customers."
"Ah. Yes, that would do it."
"A+ paper otherwise. That student's got promise."
 
Upvote
115 (118 / -3)
Post content hidden for low score. Show…

nuurdin

Smack-Fu Master, in training
60
I'd really like to see whether this holds up for models whose training data includes a healthy dose of Russian vernacular.
EDIT: I was being kind of flippant, but actually maybe this would be an legitimately interesting question. I do sort of wonder whether they would get the same results testing across languages. Russian is an interesting test case, because some of the boundaries around what might be considered "polite" speech are a little different (at least, such was the manner of speech in the post-Soviet country in which I lived many years ago).
And then I think also of testing this out in a language like Korean, which has some layers of formality in terms of verb endings and pronouns -- I mean, this is actually kind of interesting to think about, especially when you consider the gobbledygook that most training data actually consist of...
 
Last edited:
Upvote
40 (40 / 0)

Hoptimist

Ars Scholae Palatinae
712
Subscriptor++
That smiley faced image appears to show a computer not connected to the Internet. I think we have a correlation. (and what a pristine clean looking classic Mac that is)

I just asked ChatGPT to insult me - very clearly not a human. Note - I don't pay for AI, I just occasionally waste their resources.
 
Upvote
14 (16 / -2)

J.C. Helios

Ars Scholae Palatinae
989
Providing actual examples of a user’s past posts or retrieving relevant context consistently made AI text harder to distinguish from human writing, while sophisticated approaches like giving the AI a description of the user’s personality and fine-tuning the model produced negligible or adverse effects on realism.

??? What does this mean? Isn't "retrieving relevant context" table-stakes for having a conversation?
 
Upvote
18 (20 / -2)

FangsFirst

Ars Centurion
215
Subscriptor++
As someone who tries desperately to avoid toxicity (due to the anxiety it causes me—not that I don't get anxious about all interactions…), let me unleash things for just a moment:

As someone who also took up em-dashes with the ease of starting to use compose keys a few years ago, I'm fucking sick and tired of learning everything I do is what LLMs are tuned to do. This is fucking bullshit.
 
Upvote
87 (89 / -2)

telenoar

Ars Centurion
289
Subscriptor
Hmmmm… duh? Aren't most LLMs (definitely ChatGPT) trained and augmented to be more helpful and nicer than the typical response you'd get online? That's the whole point of why people veer towards AI chatbots instead of asking questions in online forums — (perceived) helpfulness (and speed).

After all, when you post a question on Reddit you know what you'll get: One third snark, with the wittiest responses earning top votes; one third helpful/empathetic responses; and one third tell you that you're the worst/dumbest/weirdest person ever to even pose such a question. Sort of doesn't matter if it's a question about coding, interior design, or AITAH.
 
Upvote
70 (70 / 0)

Fearknot

Ars Scholae Palatinae
1,341
While researchers keep trying to make AI models sound more human, actual humans on social media keep proving that authenticity often means being messy, contradictory, and occasionally unpleasant. This doesn’t mean that an AI model can’t potentially simulate that output, only that it’s much more difficult than researchers expected.

Why would "sound like a human on social media" be a design goal for AI? Not only is that hardly the gold standard for desirable behavior by humans, it also doesn't sound even close to the "killer app" that makes AI worthwhile or profitable.
 
Upvote
68 (68 / 0)

TheShark

Ars Praefectus
3,114
Subscriptor
The study also revealed an unexpected finding: instruction-tuned models, which undergo additional training to follow user instructions and behave helpfully, actually perform worse at mimicking humans than their base counterparts.

Why is this unexpected? Personally, it's exactly what I would expect. I mean the next sentence literally says that they undergo training to do something which is not mimicking humans. I'm 99% sure that Anthropic and OpenAI are not including "Dude, why would you even ask that" and "Just google it!!" as good reponses in their instruction tuning datasets.
 
Upvote
51 (51 / 0)

jasonmicron

Ars Tribunus Militum
1,787
give it time, they'll make chatbots start sounding like jerks online. those social media sites need it else they'll lose their precious engagement.
Wait until they start launching campaigns to claim that Gore won in Florida, we never landed on or went to the Moon, JFK was shot by Russian agents, the Earth is flat, the Vegas Mandalay Bay shooting was a false flag, storm systems Katrina and Harvey were caused by HAARP and whatever else triggers rage responses.
 
Upvote
12 (12 / 0)