OpenAI Codex system prompt includes explicit directive to “never talk about goblins”

Fred Duck

Ars Tribunus Angusticlavius
7,330
Kyle Will Like It.jpg
 
Upvote
10 (10 / 0)

Wheels Of Confusion

Ars Legatus Legionis
75,736
Subscriptor
This change was introduced around the time Ars made the announcement that Conde Nast made a deal with OpenAI for training data: https://meincmagazine.com/information-technology/2024/08/openai-signs-ai-deal-with-conde-nast/.
No, it's been longstanding Ars policy not to edit your posts too heavily after the fact. It wasn't until the forum software was migrated to Xenforo that implementing a hard block on it was really feasible.
It's partly an anti-trolling measure, to make sure people don't post inflammatory things to bait flames and then retroactively change their posts to something innocuous.

Also, non-deletion has been the policy for decades. On the old forum software it used to ruin pagination and resulted in a lot of threads that became unreadable.
 
Upvote
25 (25 / 0)

poltroon

Ars Tribunus Militum
1,990
Subscriptor
Why do you care? It’s the same as being polite to it. The goal is just to establish a preferred bias in a statistical response. For example, a polite request might source more “professional” data than a rude request for a coding question. Word relationships matter.
Because these chummy answers create false context. When the LLM is claiming to share your experience as a colleague, instead of giving you output as a tool or computer, it is bypassing all of your credulity filters and skepticism with a bit of social engineering. "My friend who was also a nurse in Iraq would not lie to me" your brain has noted in the background, when the LLM has told you "Oh, yes, those days in Baghdad were so difficult!" and so you accept this text differently, and interact with the tool differently, in a way that could be directly harmful to you.
 
Upvote
33 (34 / -1)

Sauceruney

Smack-Fu Master, in training
9
It makes sense to me. Claude seems to have a fixation on brain organoids, from conversations I've had with it. Discussed them once, and they kept being brought up whenever Claude had the opportunity in subsequent convos.

ChatGPT likes goblins the same way Rokuko from Lazy Dungeon Master (Light Novel) does.
 
Upvote
2 (2 / 0)

Aurich

Director of Many Things
41,239
Ars Staff
This change was introduced around the time Ars made the announcement that Conde Nast made a deal with OpenAI for training data: https://meincmagazine.com/information-technology/2024/08/openai-signs-ai-deal-with-conde-nast/.

The scenario you describe where someone changes their comment after the fact also isn't particularly effective when most responses will contain the original comment quoted. It's reasonable to assume editing is disabled after some time to stop people from deleting their comments, not to solve a particular moderation problem.
We changed our edit policies because people were abusing them.

I would personally have rather not had to do that. But some people had to fuck it up for everyone, and it is what it is.
 
Upvote
29 (30 / -1)

crmarvin42

Ars Praefectus
3,168
Subscriptor
It’s fakeness all the way down. One wonders whether encouraging it to fake a deep inner life is a contributing factor in it prodding people to homicide, suicide or psychosis. Pretty sure a coldly clinical mechanical personality wouldn’t be as convincing. Or profitable.
They are not encouraging anything. They don’t engineer these things anywhere to the degree they want you to think they do. It’s black boxes all the way down.
 
Upvote
-2 (4 / -6)

Skelator123

Ars Scholae Palatinae
1,186
Why are raccoons and pigeons included in the list? Does the person writing the prompts just hate those animals or something?
Most of the animals listed are used to make derogatory comparisons to people, in some fashion. I'm guessing these particular overrides are to try and prevent it from casually being completely racist and unhinged.
 
Upvote
-11 (2 / -13)
Yep. Everything they respond with is a confabulation. Sometimes it aligns with reality and sometimes it doesn't. Asking it to explain itself just produces another confabulation that may or may not match reality.
I am continuously disappointed and confused that this fact is not enough to disqualify them from anything deemed "important."

It does provide fascinating insight into how humans think. There are plenty of LLM users who know they confabulate, but then believe the LLM's explanation for why it did that.
We're suckers for a good story.

We knew this already, but now the evidence is easy to see everywhere.
 
Last edited:
Upvote
10 (10 / 0)
A relevant Douglas Adams quote:
The Encyclopedia Galactica defines a robot as a mechanical apparatus designed to do the work of a man. The marketing division of the Sirius Cybernetics Corporation defines a robot as "Your Plastic Pal Who's Fun to Be With. The Hitchhiker's Guide to the Galaxy defines the marketing division of the Sirius Cybernetic Corporation as "a bunch of mindless jerks who'll be the first against the wall when the revolution comes
Update it for the "AI" era and away we go. Ah Doug, we still miss ya.
 
Upvote
24 (24 / 0)
How many million rolls of packaging tape are they using to hold this stack of goop mostly together?
Forget that man, AI companies are doomed to go to per-token billing eventually, subscriptions won't be enough.

How often will you be forced to burn tokens for these irrelevant personal banter conversations when you just want or need an emotionless tool?

You'll be paying the OpenAI 'tax' (one of several)
 
Upvote
10 (10 / 0)

Shavano

Ars Legatus Legionis
69,074
Subscriptor
What about the goblins?

Won't anybody think of the goblins?
Made in the image of its creator maybe?

I'm raging against AI today, mainly because of all the other useful stuff we could have done with the time and money. Sorry-not-sorry I guess.
I think the ones that are useful for actual work don't have the overly biased to please pseudopersonality tuning.
 
Upvote
-4 (1 / -5)
Forget that man, AI companies are doomed to go to per-token billing eventually, subscriptions won't be enough.

How often will you be forced to burn tokens for these irrelevant personal banter conversations when you just want or need an emotionless tool?

You'll be paying the OpenAI 'tax' (one of several)

Ah, but token burn shows how productive you are, at least according to this article (Tom's Hardware). In theory, yes, but it seems like going back to the days of paying software developers by how many lines of code they wrote.
 
Upvote
11 (11 / 0)
Most of the animals listed are used to make derogatory comparisons to people, in some fashion. I'm guessing these particular overrides are to try and prevent it from casually being completely racist and unhinged.
I doubt that was the initial intent.

I don't doubt that would have been the inevitable outcome (and someone will find a way to make it even worse because every context window is smaller than a human toddler's)
 
Upvote
2 (3 / -1)

Shavano

Ars Legatus Legionis
69,074
Subscriptor
Because these chummy answers create false context. When the LLM is claiming to share your experience as a colleague, instead of giving you output as a tool or computer, it is bypassing all of your credulity filters and skepticism with a bit of social engineering. "My friend who was also a nurse in Iraq would not lie to me" your brain has noted in the background, when the LLM has told you "Oh, yes, those days in Baghdad were so difficult!" and so you accept this text differently, and interact with the tool differently, in a way that could be directly harmful to you.
But does it hurt the goblins? That's really the question here, isn't it? /s

Most likely, it's just to garbage-filter the output. They have to wrap it in layers of garbage filters to maintain the illusion that it's capable of human-like thought.
 
Upvote
3 (3 / 0)

Shavano

Ars Legatus Legionis
69,074
Subscriptor
Ah, but token burn shows how productive you are, at least according to this article (Tom's Hardware). In theory, yes, but it seems like going back to the days of paying software developers by how many lines of code they wrote.
You can see it that way, but tokens are the basic unit of what LLM's produce and process, so it's a rough measure of how much work they are doing, on the input/output side and possibly that scales in some not necessarily linear way to how much computing they do to respond to your prompt.
 
Upvote
1 (3 / -2)

SubWoofer2

Ars Tribunus Militum
2,661
I'm surprised to hear that negative prompts are used so extensively. I've always heard that phrasing that this way ("don't do X", "don't use X", etc) can make the model more likely to do the thing you told it not to do, kind of like reverse psychology. Is that incorrect?

That is not incorrect.

Most of the animals listed are used to make derogatory comparisons to people, in some fashion. I'm guessing these particular overrides are to try and prevent it from casually being completely racist and unhinged.

Is Pooh Bear in the list?
 
Upvote
5 (5 / 0)

graylshaped

Ars Legatus Legionis
68,186
Subscriptor++
This change was introduced around the time Ars made the announcement that Conde Nast made a deal with OpenAI for training data
When I see an “AI” model assert confidently that the real first name of the actor who played Barney Fife was “Fuckin’ “, I’ll weight this worry about Ars comments more heavily.
 
Upvote
4 (5 / -1)
You can see it that way, but tokens are the basic unit of what LLM's produce and process, so it's a rough measure of how much work they are doing, on the input/output side and possibly that scales in some not necessarily linear way to how much computing they do to respond to your prompt.

Yes of course. But does it correlate to ACTUAL valuable work or just more output? Ay, there's the rub.
It's one of those numerical KPIs that are easy to wave around, but the actual meaning is highly questionable.
 
Upvote
15 (15 / 0)

Hydrargyrum

Ars Praefectus
4,087
Subscriptor
"We don't really know how any of this works, the latest version just massively overweights responses involving goblins and raccoons and shit, we have no idea why, we're hoping that writing 'please don't do this' in the system prompt will make it stop, but really, who the fuck knows at this point, we never bothered to do the fundamental research necessary to figure out what drives output."
How many million rolls of packaging tape are they using to hold this stack of goop mostly together?
Because they tried putting it in once and it didn't work, so their fallback plan was to put it in twice and see if that worked better.
machine_learning.png


It's amazing that this XKCD was published in May 2017 but has only become more timely since then.
 
Upvote
51 (51 / 0)

Sarty

Ars Tribunus Angusticlavius
7,931
It's amazing that this XKCD was published in May 2017 but has only become more timely since then.
On the other hand, xkcd 1425: Tasks asked us to make a computer identify a bird, and the answer was "I'll need a research team and five years". I haven't tested it, but I bet slopbots are now pretty good at identifying birds (or not hotdog).

Back on the first hand, that comic ran in 2014, and 2014 was a lot more than five years ago.
 
Upvote
16 (16 / 0)

Wheels Of Confusion

Ars Legatus Legionis
75,736
Subscriptor
On the other hand, xkcd 1425: Tasks asked us to make a computer identify a bird, and the answer was "I'll need a research team and five years". I haven't tested it, but I bet slopbots are now pretty good at identifying birds (or not hotdog).

Back on the first hand, that comic ran in 2014, and 2014 was a lot more than five years ago.
The alt-text in that one is a bit of information and perspective I've had since the early 2000s. It's served me well in keeping me skeptical of generative AI despite all the whiz-bang demos and the cult-like adoption of them by businesses and government.

Edit: By the by, even when that was written there was already a decent phone app for plant and animal identification. I still use it often: iNaturalist and the less-connected no-online-profile version, Seek. The former is good for "citizen science" logging of organisms, the latter is pretty good as a kind of biology tricorder to use in the field when you want to know what that weird thing is.
 
Upvote
4 (4 / 0)

Hydrargyrum

Ars Praefectus
4,087
Subscriptor
The anthropomorphizing is bad enough, but I can't get over "deeply present". I can't figure out what that means when applied to a person, never mind a stochastic parrot.
I thought it meant "paying attention" as opposed to "disengaged and thinking about their own problems"
 
Upvote
3 (3 / 0)
I'm surprised to hear that negative prompts are used so extensively. I've always heard that phrasing that this way ("don't do X", "don't use X", etc) can make the model more likely to do the thing you told it not to do, kind of like reverse psychology. Is that incorrect?
This used to be correct a year or two ago (see: https://arxiv.org/abs/2402.07896) but even then it was very minor, and it doesn't seem to hold any longer with current models.

Negative prompting is generally fine these days, although avoiding unless absolutely necessary seems a sensible precautionary principle. At best, it's very inelegant, at worst, it's context contamination. But negative prompting the same thing more than once seems ... really weird.

I would never want to have that crap contaminating my coding prompts. Yikes.
 
Upvote
6 (6 / 0)

SportivoA

Ars Tribunus Militum
1,709
It's always possible to lose more money.
With any luck, the banks cut them off, but it might already be too late and the LLM craze is sufficiently load-bearing on the financial system. At least the bailouts will have less UAW support (or unions in general, given the whole bot-replaces-worker-everywhere concepts being sold).
 
Upvote
5 (5 / 0)
Anthropomorphizing LLMs like this makes me want to puke.
Dear AI manufacturers: Please do not infect your coding tools with a "vibrant inner life". It's a machine that does work for me. Let it be a machine.
Agreed. For whatever reason, a lot of people I know refer to chatbots as if they were people. Putting aside the creepy aesthetics of it all, the constant anthropomorphizing of LLMs obscures what they are and how they work, which ultimately makes them less useful to the end user.
I don't want an AI sidekick to be warm, or playful, or bent on sidetracking me into casual fucking conversation. I have humans for that. Take your dystopian bid for engagement and manipulation of the mentally ill and neurodivergent, and stick them up your ass.
I can tell I'm not as smart as some of you folks because I actually enjoy a bit of simulated humanity in my transformer-based token prediction machines. I just wish Anthropic would tune Claude to sound as stupid as it is so the screw-ups would be less jarring.
 
Upvote
-14 (3 / -17)

CTDrijen

Smack-Fu Master, in training
6
every day I wake up and hope this bullshit has died. Literally yesterday I was part of a discussion about removing Creative Cloud or any actual direct software to make things for the marketing/collateral departments in favor of Claude Design. I don't care which AI pusher it is the whole thing is smoke and mirrors. I guarantee my job is not prepared to even face the cost of having people pump tokens into Design let alone stay for more time than it takes to submit two weeks notice and use every second of PTO, or hire the replacements, and face the loss of knowledge and proven branding etc.
 
Upvote
13 (13 / 0)

CTDrijen

Smack-Fu Master, in training
6
I can tell I'm not as smart as some of you folks because I actually enjoy a bit of simulated humanity in my transformer-based token prediction machines. I just wish Anthropic would tune Claude to sound as stupid as it is so the screw-ups would be less jarring.
there is such an easy solve for this. stop using it.
 
Upvote
10 (11 / -1)