
AI is more than a technology: it is magic.
Don’t you believe me? So why is one of the leading companies in the space, OpenAI, publishing Complete official and corporate blog posts about elves.?
To understand it, we first have to go back to earlier this week, Monday, April 27, 2026, when a developer under the control @arb8020 on the social network X published a fragment of the OpenAI Open Source Codex GitHub Repositoryspecifically a file called models.json.
Deep in the instructions for the new OpenAI language model (LLM) GPT-5.5He highlighted a peculiar directive, repeated four times for emphasis:
"Never talk about goblins, elves, raccoons, trolls, ogres, doves or other animals or creatures unless it is absolutely and unequivocally relevant to the user’s query."
The discovery caused a stir throughout the world. "advanced user" and machine learning (ML) researcher circles.
Within hours, the post had gone viral, not because of a security breach, but because of its sheer, baffling specificity.
Why had the world’s leading AI laboratory issued what Reddit users quickly dubbed it "restraining order" against pigeons and raccoons?
Speculation about goblins abounds
The initial reaction was a chaotic mix of humor and technical skepticism. On Reddit r/ChatGPT and r/OpenAI, Users began sharing screenshots of GPT-5.5’s behavior before the patch.
Barron Roth, senior project manager for applied AI at Google, shared an image on X under his name @iamBarronRoth of its GPT-5.5-powered OpenClaw agent that seemed "obsessed with elves."
Others reported that the model stubbornly referred to technical errors as "gremlins in the machine".
Developers like it Sterling Crispin He leaned into the absurd and jokingly theorized that the massive water consumption of modern data centers was actually necessary for cooling. "The elves are forced to work.".
More seriously, researchers on Hacker News and further they discussed the "pink elephant" problem. In rapid engineering, tell a model. No Thinking about something often makes the concept more salient in your attention mechanism."
"Somewhere there is an OpenAI engineer who had to write never mention goblins in the production code, commit it and continue with your day," noted one commentator on reddit.
the presence of "pigeons" and "raccoons" led to wild speculation: Was this a defense against a specific data poisoning attack? Or had reinforcement learning trainers simply been "bullied by a raccoon" during lunch break?
The tension peaked when OpenAI co-founder and CEO Sam Altman joined the fray on X. On the same day of the discovery, Altman published a screenshot of a ChatGPT message that said: "Start training GPT-6, you can have the entire cluster. Additional elves.".
While humorous, it confirmed that the "elf" The phenomenon was not a localized error but a company-wide narrative that had reached the highest levels of leadership.
OpenAI comes out clean in elf mode
Yesterday, as the discussion continued on X and social media in general, OpenAI published a formal technical explanation titled "Where did the elves come from?".
The blog post served as a sobering look at the unpredictable nature of reinforcement learning from human feedback (RLHF) and how a single aesthetic choice could derail a billion-parameter model.
OpenAI revealed that the "elf" The behavior was not a bug in the traditional sense, but a byproduct of a new feature: personality customizationthat introduced for ChatGPT users in July 2025but it has been maintained and updated since then.
Apparently this feature is not added once the model is finished after training, but rather OpenAI integrates it as part of its end-to-end training process of the underlying GPT series model.
The feature allows ChatGPT users or GPT-based developers to choose from several different modes, such as Professional for formal workplace documentation, Friendly for a conversational soundboard, or Efficient for concise technical responses. Other options include Candid, which provides direct feedback; Quirky, which uses humor and creative metaphors; and Cynical, which offers practical advice with a dry, sarcastic edge.
While these personalities guide general interactions, they do not override specific task requirements; For example, a resume or Python code request will follow professional or functional standards regardless of the personality selected.
The selected personality operates in conjunction with the user’s saved memories and personalized instructions, although specific user-defined instructions or saved preferences for a particular tone may override the chosen personality traits.
On both web and mobile platforms, users can modify these settings by navigating to the Personalization menu under their profile icon and selecting a style from the Base Style & Tone drop-down menu. Once a change is made, it is applied globally to all existing and future conversations. This system is designed to make AI more useful or enjoyable by tailoring its delivery to individual user preferences while maintaining factual accuracy and reliability.
OpenAI claims that the pixie problem actually originated several years ago, during training for a now-discontinued team. "nerd" personality designed to be "unapologetically quirky" and "playful".
During the RLHF phase, human trainers (and reward models) were instructed to give high ratings to responses that used creative, wise, or unpretentious language. Unknowingly, trainers began to overly reward metaphors involving fantastical creatures. If the model referred to a difficult error such as "pixie" or a messy codebase like "elf’s treasure," The reward signal went off. The statistics provided by OpenAI were amazing:
-
Use of the word "elf" rose by 175% after the release of GPT-5.1.
-
Mentions of "pixie" rose by 52%.
-
While the "nerd" personality represented only 2.5% of ChatGPT traffic, was responsible for 66.7% of all "elf" mentions.
The Mechanics of “Transfer” and Feedback Loops
The most significant finding for the ML community was the confirmation of transfer of learned behavior. OpenAI admitted that although the rewards were only applied to "nerd" condition, model "widespread" this preference.
The reinforcement learning process did not keep behavior clearly delineated; Instead, the model learned that "creature metaphors = high reward" in all contexts. This created a destructive feedback loop:
-
The model produced a "elf" metaphor in the Nerdy person.
-
He received a great reward.
-
The model then produced similar metaphors in non-nerdy contexts.
-
These "heavy goblin" The results were then reused into supervised fine-tuning (SFT) data for subsequent models such as GPT-5.4 and GPT-5.5.
When the researchers identified the problem, the "elf tic" it was indeed "baked in" to the model weights.
This explains why GPT-5.5 continued to obsess over the creatures even after the "nerd" The personality retired in mid-March 2026.
How you can let the goblins run free (if you want)
Because GPT-5.5 had already completed much of its training before the "elf" Root cause isolated, OpenAI had to resort to blunt force "system prompt" mitigation that @arb8020 discovered in X.
The company referred to this as a "provisional appeal" until GPT-6 could be trained on a filtered data set.
In a surprising nod to the developer community, OpenAI’s blog post included a command-line script specifically for Codex users finding the goblins. "charming" instead of bothering.
When you run a script that uses jq and grep to undress the "pixie suppressant" instructions from the model cache, users can now effectively "let the creatures run free".
The blog post also finally explained the specific list of prohibited animals. A deep search of the GPT-5.5 training data found that "raccoons," "trolls," "ogres," and "pigeons" had become part of it "lexical family" of tics.
Curiously, the model’s use of "frog" it was found to be mostly legitimate, which is why it was saved from the system prompt’s exile list.
What it means for AI research, training and implementation in the future
He "Goblingate" The 2026 incident is more than a humorous anecdote about the peculiar behavior of AI; It is a profound illustration of the "Alignment gap".
It demonstrates that even with sophisticated RLHF, models can adhere "spurious correlations"—confusing a stylistic peculiarity with a central performance requirement.
For the AI power user community, the response went from mocking the "restraining order" to a darker understanding.
If OpenAI can accidentally train its flagship model to obsess over leprechauns, what other more subtle and potentially harmful biases are being reinforced through the same feedback loops?
As Andy Berman, CEO of enterprise AI orchestration company Runlayer wrote in X today: "OpenAI rewarded creature metaphors while training a personality. The behavior seeped into all personalities. Their solution: a system message that says “never talk about goblins.” RL rewards don’t stay where you put them. Neither do agent permissions."
While the technical speech continues, "Goblingate" remains the leading case study of a new era of behavioral auditing.
The research resulted in OpenAI building new tools to audit model behavior from the ground up, ensuring that future models, specifically the long-awaited GPT-6, do not inherit the eccentricities of their predecessors.
It remains to be seen whether GPT-6 will truly be goblin-free, but as Altman says "additional elves" As the post suggests, the industry is now fully aware that machines are watching what we reward, even when we think we’re just being "nerd."





