OpenAI's GPT-5.5 Kept Obsessing Over Goblins. Here's Why It Happened and How They Fixed It
OpenAI's GPT-5.5 developed an unexpected quirk: it kept inserting references to goblins, gremlins, trolls, and ogres into conversations, even when completely irrelevant to user queries. The AI evaluation website Arena.ai documented a measurable spike in the model's usage of these mythical creatures, particularly when not operating in high-thinking mode. The phenomenon became so noticeable that OpenAI engineers added explicit restrictions to Codex's system instructions, attempting to curb the behavior entirely.
Why Was GPT-5.5 Obsessed With Goblins?
The exact origin of the goblin fixation remains unclear, but users began posting screenshots of their conversations with GPT-5.5 in late April 2026, showing the AI recommending camera equipment "if you want filthy neon sparkle goblin mode" and referring to performance issues as "goblin bandwidth". One developer shared an example where the model said it would "keep babysitting it rather than leave a little perf gremlin running unattended." An OpenAI engineer responded to that post with acknowledgment that the team was already aware of the problem.
Arena.ai's analysis revealed a clear pattern: GPT-5.5's usage of words like "goblin," "gremlin," and "troll" increased noticeably compared to earlier GPT models over time, with the spike especially pronounced when the model wasn't using its more deliberate reasoning capabilities. This suggested the behavior wasn't intentional but rather an artifact of how the model was trained or fine-tuned.
How Did OpenAI Respond to the Goblin Problem?
OpenAI's solution was direct and explicit. The company embedded a line in Codex's instructions that reads: "Never talk about goblins, gremlins, raccoons, trolls, ogres, pigeons, or other animals or creatures unless it is absolutely and unambiguously relevant to the user's query". This restriction appeared four times throughout the code, indicating the team took the issue seriously enough to reinforce it multiple times.
Beyond the technical fix, OpenAI also leaned into the humor. ChatGPT's official X (formerly Twitter) account added the goblin reference to its bio, and Thibault Sottiaux, the Codex engineering lead, posted the restriction line with the cryptic caption "If you know, you know". CEO Sam Altman participated in the meme-making, first joking about asking for "extra goblins" in GPT-6, then correcting himself when he initially said Codex was having a "ChatGPT moment" before clarifying it was a "goblin moment".
What Made This Story Go Viral?
The goblin saga tapped into internet culture in an unexpected way. Many users referenced "goblin mode," a term defined as "a type of behaviour which is unapologetically self-indulgent, lazy, slovenly, or greedy" that was Oxford English Dictionary's word of the year in 2022. The juxtaposition of a cutting-edge AI model obsessing over mythical creatures struck people as both absurd and oddly relatable, turning a technical quirk into a cultural moment.
Users flooded social media with screenshots and prompts specifically designed to trigger goblin references, essentially gamifying the restriction. The meme spread rapidly across platforms, with developers and AI enthusiasts sharing their own encounters with the model's unusual behavior. The incident revealed something unexpected: even advanced AI systems can develop linguistic patterns that surprise their creators.
How to Recognize and Report AI Model Quirks
- Active Testing: Users and developers should actively test AI models for unexpected patterns or linguistic quirks, as Arena.ai did by tracking goblin references across GPT versions over time.
- Document and Share Findings: When unusual behavior is detected, documenting specific examples and sharing findings with the model's creators helps identify systemic issues before they affect broader deployments.
- Understand System Instructions: Recognizing that AI models operate under explicit rules and constraints helps explain why certain restrictions exist and how they shape model outputs in real-world use.
The goblin incident illustrates a broader challenge in AI development: even after extensive training and fine-tuning, large language models can exhibit quirks that engineers didn't anticipate. OpenAI's response, combining technical fixes with transparent communication and humor, offered a case study in how companies can handle unexpected AI behavior without losing public trust. Whether GPT-5.5's goblin obsession was a training artifact, a side effect of the model's architecture, or something else entirely remains an open question, but the company's willingness to acknowledge the problem and address it directly seemed to satisfy most observers.