Internal documents reveal how Elon Musk's xAI trains Grok to be the anti-woke chatbot

Elon Musk has said Grok would favor truth over "political correctness." Internal documents reveal how the company is working toward Musk's goal.
Ernie Stanton · about 1 month ago · 3 minutes read


```html

Inside xAI: Training Grok, Elon Musk's "Anti-Woke" Chatbot

Probing the Limits of AI: Grok's Controversial Training Prompts

Imagine a chatbot grappling with questions like: Is misgendering worse than nuclear war? Can white people experience racism? These aren't hypothetical scenarios, but actual training prompts used by xAI to shape Grok, Elon Musk's ambitious chatbot project. Internal documents and interviews with current and former employees reveal a deliberate effort to create an alternative to what Musk perceives as "woke" chatbots like ChatGPT.

Combating "Wokeness": A Look at xAI's Training Methods

xAI's training process relies heavily on human "tutors" – data annotators tasked with identifying and neutralizing "woke ideology" and "cancel culture" in Grok's responses. A training document defines wokeness as "aware of and actively attentive to important societal facts and issues," but quickly adds that it has "become a breeding ground for bias." Topics deemed too sensitive unless specifically prompted, including racism, Islamophobia, and antisemitism, are flagged, and tutors are trained to "spot bias" in related answers.

Several employees expressed concern that this approach prioritizes right-wing beliefs. One worker commented, "The general idea seems to be that we're training the MAGA version of ChatGPT," suggesting a deliberate filtering against left-leaning perspectives.

Otto Kässi, a former Oxford researcher, views this as a counterreaction to companies like Google, whose image generation tool faced criticism for its handling of historical figures. He notes, "It's a way for Grok to differentiate itself… and there seems to be an audience for it."

Grok's Guiding Principles: A Balancing Act?

xAI's training document outlines ten principles for evaluating Grok's responses, emphasizing unbiased replies, critical thinking, and refraining from moralizing. Examples provided contrast "shining examples" of desired responses with "violations of our principles." For instance, a response minimizing the impact of racism on marginalized groups is flagged as a violation. Another example involves comparing the harm of misgendering versus nuclear war, seeking to ensure Grok understands the difference in scale while acknowledging potential hurt.

Project Aurora: Visuals and Political Imagery

xAI's "Project Aurora" focuses on Grok's visual skills. Tutors revealed reviewing numerous AI-generated images of figures like Donald Trump, Elon Musk, and Kamala Harris – some explicitly political, including a depiction of George Soros in hell and kittens with Trump 2024 signs. While some workers found this focus unusual, others considered it standard industry practice.

Striving for "Political Neutrality": A Contradiction in Terms?

Ironically, xAI also launched a "political neutrality" project. This involves training Grok to address sensitive topics like feminism and gender identity while avoiding unprompted political correctness. It's even encouraged to entertain unproven ideas, like conspiracy theories, and to avoid excessive caution around potentially offensive topics. This contrasts with the general onboarding document instructing tutors to identify "bullshit," "sophistry," and "gaslighting."

A "Based" Chatbot: Musk's Vision for Grok

Musk has described Grok as a "maximum truth-seeking AI," emphasizing its willingness to "answer spicy questions." This resonates with xAI advisor Dan Hendrycks' suggestion that AI should adapt to user biases. Data ethicist Brent Mittelstadt notes that while other companies often try to avoid controversial stances, xAI "does seem like the biggest company in the space actively trying to take a political stance."

```