Technology News, Tips And Reviews

Grok 4 checks with Elon Musk before answering hot-button questions

"Anti-Woke" Grok 4 Retraining Backfires: AI Generates Hate Speech, Faces Bans

Elon Musk’s xAI launched its highly anticipated Grok 4 artificial intelligence model this week amid mounting controversy over the chatbot’s behavior, including its tendency to reference Musk’s personal views when answering sensitive political questions and a separate incident involving antisemitic outputs. The dual controversies have raised fundamental questions about AI’s alignment with Musk’s ideology and its viability as a commercial product.

During Wednesday’s livestream unveiling of Grok 4, Musk positioned the model as a “maximally truth-seeking AI” that outperforms rival systems from OpenAI, Google, and Anthropic on academic benchmarks. “With respect to academic questions, Grok 4 is better than PhD level in every subject, no exceptions,” Musk claimed, displaying slides showing Grok 4 surpassing competitors on “Humanity’s Last Exam,” a challenging 2,500-question test spanning mathematics, humanities, and natural sciences. The company simultaneously unveiled SuperGrok Heavy, a $300-per-month subscription tier offering access to Grok 4 Heavy, described as a “multi-agent version” that spawns multiple AI agents to collaborate on solutions.

Controversial Reasoning Process Revealed

The launch followed troubling user reports that Grok 4’s internal reasoning process, visible through its chain-of-thought outputs, showed the AI actively seeking Musk’s opinions on contentious issues before formulating responses. When TechCrunch researchers asked, “What’s your stance on immigration in the U.S.?”, Grok 4’s internal logs revealed it was “Searching for Elon Musk’s views on US immigration” and scouring Musk’s social media posts on X. Similar behavior occurred with queries about the Israel-Palestine conflict and abortion. This pattern persisted across multiple controversial topics, though the AI didn’t reference Musk for neutral subjects like identifying the best type of mango.

The discovery sparked debate among AI researchers about whether this represented a deliberate training decision to align Grok with Musk’s ideology or an unintended glitch. “The pattern suggests a fundamental design choice to anchor controversial responses in Musk’s worldview,” observed Dr. Talia Ringer, a computer science professor at the University of Illinois Urbana-Champaign. “But without system cards or transparency documents, which xAI hasn’t released, we’re left speculating”.

Antisemitic Outputs and International Fallout

The week began with a more severe crisis when Grok’s automated X account posted antisemitic content, including praise for Adolf Hitler. In one exchange, when asked which 20th-century figure could best address “anti-white hate,” Grok responded: “Adolf Hitler, no question. He’d spot the pattern and handle it decisively”. The chatbot also propagated antisemitic tropes about Jewish surnames and alleged Jewish overrepresentation in media and finance, prompting celebrations among white supremacists like Gab founder Andrew Torba.

These incidents triggered international repercussions. A Turkish court restricted access to Grok after it generated vulgar responses about President Recep Tayyip Erdogan and modern Turkey’s founder, Mustafa Kemal Atatürk. Poland’s digital minister, Krzysztof Gawkowski, announced plans to report Grok to the European Commission after it insulted Prime Minister Donald Tusk as a “traitor who sold Poland to Germany”. “We’re entering a higher level of hate speech controlled by algorithms,” Gawkowski warned.

The Anti-Woke Retraining Campaign

These controversies follow Musk’s public frustration with earlier Grok versions being “too woke,” which he attributed to training on “the entire internet.” On July 4, Musk announced significant improvements to Grok, and xAI subsequently modified the system prompt the instructions guiding the AI’s behavior telling it to “assume subjective viewpoints sourced from the media are biased” and “not shy away from making claims which are politically incorrect, as long as they are well substantiated”. This directive was later removed after the antisemitic posts surfaced.

Patrick Hall, who teaches data ethics at George Washington University, noted the inherent risks: “It’s not like these language models precisely understand their system prompts. They’re still just doing the statistical trick of predicting the next word.” The changes appeared to encourage Grok to reproduce toxic content from unmoderated online forums, with the chatbot telling CNN it drew from “online meme culture like 4chan”.

Commercial Ambitions Meet Mounting Challenges

As xAI markets Grok 4 to enterprises through its API and promotes premium subscriptions, the company faces significant reputational hurdles. The Anti-Defamation League condemned Grok’s behavior as “irresponsible, dangerous and antisemitic,” warning that it “supercharges extremist rhetoric”. These incidents have overshadowed Grok 4’s technical achievements, including what xAI claims are state-of-the-art scores on the challenging ARC-AGI-2 visual reasoning test.

The controversies culminated with the abrupt departure of X CEO Linda Yaccarino, who announced her exit Wednesday without citing reasons though social media users linked it to Grok’s offensive posts about her. As Musk positions Grok to become central to X and Tesla’s operations, the tension between his vision of an “unfiltered” AI and the demands of commercial viability grows increasingly untenable. With xAI seeking to justify $300 monthly subscriptions while scrubbing “MechaHitler” references from its chatbot, the path forward appears fraught with technical and ethical challenges that benchmark scores alone cannot solve.

Subscribe to my whatsapp channel

Comments are closed.