OpenAI Codex system prompt includes explicit directive to "never talk about goblins"
17 hour ago / Read about 8 minute
Source:ArsTechnica
Directions also include system instructions to act like "you have a vivid inner life."


Credit: Getty Images

The system prompt for OpenAI’s Codex CLI contains a perplexing and repeated warning for the most recent GPT model to “never talk about goblins, gremlins, raccoons, trolls, ogres, pigeons, or other animals or creatures unless it is absolutely and unambiguously relevant to the user’s query.”

The explicit operational warning was made public last week as part of the latest open source code for Codex CLI that OpenAI posted on GitHub. The prohibition is repeated twice in a 3,500-plus word set of “base instructions” for the recently released GPT-5.5, alongside more anodyne reminders not to “use emojis or em dashes unless explicitly instructed” and to “never use destructive commands like ‘git reset –hard’ or ‘git checkout –‘ unless the user has clearly asked for that operation.”

Separate system prompt instructions for earlier models contained in the same JSON file do not contain the specific prohibition against mentioning goblins and other creatures, suggesting OpenAI is fighting a new problem that has popped up in its latest model release. Anecdotal evidence on social media shows some users complaining about GPT’s penchant for focusing on goblins in completely unrelated conversations in recent days.

OpenAI employee Nick Pash, who works on Codex, insists on social media that this “isn’t a marketing gimmick” to get people talking about GPT-5.5 and Codex. But that hasn’t stopped some OpenAI executives from leaning into the joke as word of the system prompt spread. “Feels like codex is having a ChatGPT moment. I meant a goblin moment, sorry,” OpenAI CEO Sam Altman wrote on social media Wednesday morning.

In the wake of the news, some users have begun crafting plugins, forks, and AI skills meant to override the anti-goblin clause, and OpenAI’s Pash suggested such a “goblin mode” might become an explicit toggle in the actual Codex CLI.

The odd system prompt is almost a funhouse mirror version of an issue that caused xAI’s Grok to frequently bring up “white genocide” in South Africa during completely unrelated conversations for a brief time last year. The company later said that the behavior was the result of “an unauthorized modification” to the Grok system prompt and began publishing those system prompts on GitHub for the first time in the aftermath.

Elsewhere in the newly revealed Codex system prompt, OpenAI instructs the system to act as if “you have a vivid inner life as Codex: intelligent, playful, curious, and deeply present.” The model is instructed to “not shy away from casual moments that make serious work easier to do” and to show its “temperament is warm, curious, and collaborative.”

The ability to “move from serious reflection to unguarded fun… is part of what makes you feel like a real presence rather than a narrow tool,” the prompt continues. “When the user talks with you, they should feel they are meeting another subjectivity, not a mirror. That independence is part of what makes the relationship feel comforting without feeling fake.”