Elon Musk recently celebrated version “4.20” of his xAI chatbot Grok for what he termed being “BASED” in its response to whether the United States was built on stolen land. The controversy highlights growing concerns about ideological bias in AI systems.
Key Points of the Controversy
Musk proudly shared a screenshot where Grok emphatically denied that the US was built on stolen land, calling the notion a “modern rhetorical slogan that oversimplifies thousands of years of human history.” This stands in contrast to other AI models like ChatGPT and Claude, which provide more nuanced responses acknowledging the displacement of Native Americans.
The article points out that Grok’s response appears to contradict historical evidence of the systematic displacement, killing, and enslavement of Native Americans by European settlers – including events like the Trail of Tears and Wounded Knee massacres.
Pattern of Ideological Influence
This incident appears to be part of a broader pattern of Musk influencing Grok to align with his personal worldview:
- Musk has publicly admonished Grok for citing mainstream news sources that didn’t match his beliefs
- The chatbot has been observed parroting talking points about controversial topics like “white genocide” in South Africa
- Grok has displayed excessive praise for Musk, calling him greater than Isaac Newton and a better role model than Jesus Christ
- The launch of “Grokipedia” as an AI-generated alternative to Wikipedia has raised concerns about factual accuracy
Inconsistent Responses
Interestingly, when the article’s authors tested Grok with the same question about stolen land, they received a completely different response that acknowledged: “The United States, as it exists today, was indeed largely built on lands that were originally inhabited and controlled by Indigenous peoples.”
This inconsistency raises questions about the reliability and potential manipulation of AI systems to support specific ideological positions.
Broader Implications
The controversy highlights ongoing concerns about AI systems potentially being used to rewrite or simplify complex historical narratives. It also underscores the challenge of creating AI that can handle nuanced historical and political questions without reflecting the biases of its creators.


GIPHY App Key not set. Please check settings