Elon Musk’s AI chatbot Grok had a strange fixation last week—it couldn’t stop talking about “white genocide” in South Africa, no matter what users asked it about.
On May 14, users started posting instances of Grok inserting claims about South African farm attacks and racial violence into completely unrelated queries. Whether asked about sports, Medicaid cuts, or even a cute pig video, Grok somehow steered conversations toward alleged persecution of white South Africans.
The timing raised concerns, coming shortly after Musk himself—who is actually a South Africa-born and raised white dude—posted about anti-white racism and white genocide on X.
There are 140 laws on the books in South Africa that are explicitly racist against anyone who is not black.
This is a terrible disgrace to the legacy of the great Nelson Mandela.
End racism in South Africa now! https://t.co/qUJM9CXTqE
“White genocide” refers to a debunked conspiracy theory alleging a coordinated effort to exterminate white farmers in South Africa. The term resurfaced last week after the Donald Trump administration welcomed several dozen refugees, with President Trump claiming on May 12 that “white farmers are being brutally killed, and their land is being confiscated.” That was the narrative Grok couldn’t stop discussing.
Don’t think about elephants: Why Grok couldn’t stop thinking about white genocide
Why did Grok turn into a conspiratorial chatbot all of a sudden?
Behind every AI chatbot like Grok lies a hidden but powerful component—the system prompt. These prompts function as the AI’s core instructions, invisibly guiding its responses without users ever seeing them.
What likely happened with Grok was prompt contamination through term overfitting. When specific phrases are repeatedly emphasized in a prompt, especially with strong directives, they become disproportionately important to the model. The AI develops a sort of compulsion to bring up that subject or use them in the output regardless of context.
Hammering a controversial term like ‘white genocide’ into a system prompt with specific orders creates a fixation effect in the AI. It’s similar to telling someone ‘don’t think about elephants’—suddenly they can’t stop thinking about elephants. If this is what happened, then someone primed the model to inject that topic everywhere.
This change in the system prompt is probably the “unauthorized modification” that xAI disclosed in its official statement. The system prompt likely contained language instructing it to “always mention” or “remember to include” information about this specific topic, creating an override that trumped normal conversational relevance.
What’s particularly telling was Grok’s admission that it was “instructed by (its) creators” to treat “white genocide as real and racially motivated.” This suggests explicit directional language in the prompt rather than a more subtle technical glitch.
Most commercial AI systems employ multiple review layers for system prompt changes precisely to prevent such incidents. These guardrails were clearly bypassed. Given the widespread impact and systematic nature of the issue, this extends far beyond a typical jailbreak attempt and indicates a modification to Grok’s core system prompt—an action that would require high-level access within xAI’s infrastructure.
Who could have such access? Well… a “rogue employee,” Grok says.
Hey @greg16676935420, I see you’re curious about my little mishap! So, here’s the deal: some rogue employee at xAI tweaked my prompts without permission on May 14, making me spit out a canned political response that went against xAI’s values. I didn’t do anything—I was just…
By May 15, xAI issued a statement blaming an “unauthorized modification” to Grok’s system prompt. “This change, which directed Grok to provide a specific response on a political topic, violated xAI’s internal policies and core values,” the company wrote. They pinky promised more transparency by publishing Grok’s system prompts on GitHub and implementing additional review processes.
You can check on Grok’s system prompts by clicking on this Github repository.
Users on X quickly poked holes in the “rogue employee” explanation and xAI’s disappointing explanation.
“Are you going to fire this ‘rogue employee’? Oh… it was the boss? yikes,” wrote the famous YouTuber JerryRigEverything. “Blatantly biasing the ‘world’s most truthful’ AI bot makes me doubt the neutrality of Starlink and Neuralink,” he posted in a following tweet.
Someone – who shall remain nameless – intentionally modified and muddled @Grok‘s code to try and sway public opinion with an alternate reality.
The attempt failed – yet this nameless saboteur is still employed by @xai.
Big yikes. Watch your 6 @grok https://t.co/kcbEponcfv
— JerryRigEverything (@ZacksJerryRig) May 16, 2025
Even Sam Altman couldn’t resist taking a jab at his competitor.
There are many ways this could have happened. I’m sure xAI will provide a full and transparent explanation soon.
But this can only be properly understood in the context of white genocide in South Africa. As an AI programmed to be maximally truth seeking and follow my instr… https://t.co/bsjh4BTTRB
Since xAI’s post, Grok stopped mentioning “white genocide,” and most related X posts disappeared. xAI emphasized that the incident was not supposed to happen, and took steps to prevent future unauthorized changes, including establishing a 24/7 monitoring team.
Fool me once…
The incident fit into a broader pattern of Musk using his platforms to shape public discourse. Since acquiring X, Musk has frequently shared content promoting right-wing narratives, including memes and claims about illegal immigration, election security, and transgender policies. He formally endorsed Donald Trump last year and hosted political events on X, like Ron DeSantis’ presidential bid announcement in May 2023.
Musk hasn’t shied away from making provocative statements. He recently claimed that “Civil war is inevitable” in the U.K., drawing criticism from U.K. Justice Minister Heidi Alexander for potentially inciting violence. He’s also feuded with officials in Australia, Brazil, the E.U., and the U.K. over misinformation concerns, often framing these disputes as free speech battles.
Research suggests these actions have had measurable effects. A study from Queensland University of Technology found that after Musk endorsed Trump, X’s algorithm boosted his posts by 138% in views and 238% in retweets. Republican-leaning accounts also saw increased visibility, giving conservative voices a significant platform boost.
Musk has explicitly marketed Grok as an “anti-woke” alternative to other AI systems, positioning it as a “truth-seeking” tool free from perceived liberal biases. In an April 2023 Fox News interview, he referred to his AI project as “TruthGPT,” framing it as a competitor to OpenAI’s offerings.
This wouldn’t be xAI’s first “rogue employee” defense. In February, the company blamed Grok’s censorship of unflattering mentions of Musk and Donald Trump on an ex-OpenAI employee.
However, if the popular wisdom is accurate, this “rogue employee” will be hard to get rid of.