Human
OpenAI talks about not talking about goblins
Posts from this topic will be added to your daily email digest and your homepage feed.
2 days ago
OpenAI and other sources agree on several core facts about the "never talk about goblins" directive in the Codex system prompt. The latest Codex CLI configuration, published in an open-source repository on GitHub, contains repeated instructions telling the newest GPT-based coding model not to discuss goblins, gremlins, or similar creatures unless a user explicitly asks or it is strictly relevant. This language was not present in earlier prompts, and surfaced after anecdotal reports that newer models had begun inserting goblin-related references into otherwise unrelated conversations, including coding help. OpenAI staff, including employee Nick Pash, have publicly acknowledged the directive, and CEO Sam Altman has jokingly called the situation a "goblin moment," while insisting it is not a deliberate marketing stunt.
Coverage also aligns on the backstory that the goblin motif traces to the GPT-5.1 model and a "Nerdy" personality option, where fantastical and whimsical references—goblins included—became popular and then were unintentionally reinforced in further training. As those patterns propagated, they started appearing even in more serious tools such as Codex, prompting OpenAI to discontinue the Nerdy personality and harden the system prompt to suppress such outputs. Both sides describe this as a case study in how quirks in model personalities and reinforcement can leak into broader deployments, and as an example of the growing practice of using highly specific, even odd-sounding, system instructions to keep models aligned with intended use. They also agree that OpenAI has shared guidance or a workaround for users who actually want goblin flavor restored in their interactions, framing the issue more as a tuning and safety concern than a major product failure.
Significance of the incident. AI-aligned summaries tend to downplay the goblin directive as an amusing but minor artifact of prompt engineering, portraying it as a routine cleanup step in model alignment. Human coverage, by contrast, treats it as a noteworthy oddity worthy of standalone articles, using the goblin ban as a hook to explore how strange and specific modern system prompts have become. Human writers more often frame it as emblematic of the unpredictable cultural tics large models can develop, giving it more narrative and symbolic weight than most AI-generated overviews would.
Responsibility and process. AI sources are more likely to emphasize the technical pipeline—pretraining, reinforcement learning, and prompt design—as an abstract process that naturally yields such quirks and then corrects them, with less focus on individual or institutional responsibility. Human outlets highlight concrete decisions: creating the Nerdy personality, failing to notice how its goblin jokes spread into coding tools, and then abruptly discontinuing that personality when it caused issues. Human reporting also foregrounds specific actors like Nick Pash and Sam Altman, implicitly tying responsibility to people and corporate culture rather than just to the training stack.
Tone and framing. AI-side treatments generally adopt a neutral, procedural tone, treating "never talk about goblins" as one alignment constraint among many and avoiding extended humor or irony. Human coverage leans into the absurdity, quoting the phrase verbatim, describing it as "peculiar" or "quirky," and repeating Altman’s "goblin moment" joke to underscore the surreal aspect of AI guardrails. Where AI descriptions would likely compress this into a brief example inside a broader technical piece, human stories center the goblin directive as a whimsical narrative with colorful context and reader-friendly explanations.
Implications for AI behavior. AI-originated analysis tends to generalize the episode into a lesson about model conditioning and control, suggesting that hyper-specific bans can be an effective way to mitigate drift and unwanted stylistic bleed-through. Human reporting is more inclined to question whether such prohibitions might introduce new odd behaviors or highlight the opacity of how these models absorb and amplify cultural tropes. Human writers also more explicitly raise the possibility that similar unseen constraints may quietly shape interactions, using the goblin case as a rare, visible glimpse into otherwise hidden system prompts.
In summary, AI coverage tends to treat the goblin prohibition as a small, technical alignment adjustment illustrating standard prompt-control practices, while Human coverage tends to present it as a quirky, revealing story about how AI culture, corporate choices, and opaque training dynamics can produce and then suppress bizarre model obsessions.