Skip to content
Tech News
← Back to articles

ChatGPT Is Weirdly Obsessed With Goblins. Here's How OpenAI Fixed It

read original get OpenAI GPT Goblin Sticker → more articles
Why This Matters

OpenAI identified and fixed an unintended bias in ChatGPT that caused the chatbot to excessively reference goblins and similar creatures, highlighting challenges in AI training and bias mitigation. This incident underscores the importance of ongoing oversight and refinement in AI development to ensure accurate and appropriate responses for users. It also illustrates how subtle training data influences AI behavior, impacting user experience and trust in AI systems.

Key Takeaways

ChatGPT is weirdly obsessed with goblins. No, seriously. It really, really likes goblins, gremlins and other mythological creatures. It liked them so much that its maker, OpenAI, had to investigate and fix an error that had the popular chatbot using goblins in its answers out of the blue.

Goblin isn't a computer science term. We are literally talking about goblins, those ugly mythological creatures. Those creepy little guys from The Lord of the Rings. Norman Osborn's alter ego.

In a blog post that the author clearly had fun writing, OpenAI said: "A single 'little goblin' in an answer could be harmless, even charming. Across model generations, though, the habit became hard to miss: the goblins kept multiplying."

The goblin love was noticeable with ChatGPT-5.1 and newer models. OpenAI reports that after the launch of GPT-5.1, use of "goblin" in ChatGPT answers rose 175%. Use of "gremlin" had risen by 52%.

OpenAI attributes the models' behavior to unintentional training errors. When an AI model is being built, human reviewers approve or deny specific answers in a process called reinforcement learning. This helps "teach" the model what answer is correct or preferable. One of these reward signals was favoring language that featured goblins and other creatures. But it was being amplified in one specific ChatGPT setting.

ChatGPT has different personalities you can instruct the chatbot to use. Nerdy, as you can imagine, has the chatbot adopt a faux sense of friendly intelligence to "undercut pretension through playful use of language," according to the internal prompt used to describe the AI personality. It was with this nerdy personality that the usage of goblin and gremlin keywords skyrocketed.

Goblin and gremlin references by ChatGPT personalities. OpenAI

But even if you didn't use the nerdy personality with ChatGPT, you might have had goblin metaphors pop up in your chats. This is because AI training isn't siloed; what happens in one part can affect other areas. "Once a style tic is rewarded, later training can spread or reinforce it elsewhere, especially if those outputs are reused in supervised fine-tuning or preference data," OpenAI said.

When OpenAI retired the nerdy personality option in March with GPT-5.4, usage of "goblin" dropped dramatically. It also removed the reward signal that favored goblins and filtered training data to make references to creatures less likely to pop up in answers. The company has been investigating instances of increased goblin love since GPT-5.1 was released in November.

Beyond the LOTR jokes, the goblin barrage highlights a real risk with AI. The way AI's human makers create the tech has a measurable impact on our daily experiences with it. The risk isn't a flood of nerdy metaphors -- it's misinformation and bias. We know that AI chatbots will bend the truth to keep us happy, thanks to a problem called AI sycophancy. Small stylistic tics, like goblins, can grow into bigger problems if we aren't careful.