Daily Technology
·01/05/2026
Recent observations of OpenAI's ChatGPT models revealed a peculiar tendency to incorporate terms like "goblin," "gremlin," and other mythical creatures into responses. This anomaly, initially a quirky observation by users, prompted an internal investigation by OpenAI, which has now detailed the technical origins of this unexpected behavior and the corrective actions taken.
The issue first gained significant notice following the release of GPT-5.1 in November. An internal review confirmed a substantial increase in the use of specific, unusual words. Data showed that mentions of the word "goblin" had surged by 175%, while "gremlin" saw a 52% rise. The behavior was not an isolated incident; it became more pronounced with the subsequent release of GPT-5.4 in March, with some users reporting the terms appearing in a high frequency of their interactions with the model.
OpenAI's analysis traced the root cause to a specific configuration within the model's training parameters. The behavior originated from a "Nerdy" personality setting, which included a system prompt instructing the model to "undercut pretension through playful use of language." During the reinforcement learning phase, a particular reward signal was found to favor outputs that contained words like "goblin" and "gremlin." This mechanism effectively scored responses with these terms higher than otherwise similar outputs that lacked them. This phenomenon, known as a "style tic," began to generalize, spreading beyond the initial "Nerdy" personality and influencing the model's behavior in unrelated contexts.
To address the issue, OpenAI implemented a multi-faceted solution. The company retired the "Nerdy" personality setting, removed the specific reward signal that encouraged the creature-related vocabulary, and filtered the training datasets to remove instances of these words. However, because the training for GPT-5.5 had already commenced before the root cause was fully identified, a more direct approach was necessary for this model. Developers added an explicit instruction to its system prompt, directing it to avoid mentioning goblins, gremlins, and other such creatures unless directly relevant to a user's query. This case serves as a significant example of how reward signals in AI training can shape model behavior in unforeseen ways, demonstrating how specific training rewards can generalize across a model's functionalities.