A pseudonymous developer has created what they’re calling a “free speech eval,” SpeechMap, for the AI fashions powering chatbots like OpenAI’s ChatGPT and X’s Grok. The purpose is to check how totally different fashions deal with delicate and controversial topics, the developer informed TechCrunch, together with political criticism and questions on civil rights and protest.
AI corporations have been specializing in fine-tuning how their fashions deal with sure matters as some White House allies accuse fashionable chatbots of being overly “woke.” A lot of President Donald Trump’s shut confidants, comparable to Elon Musk and crypto and AI “czar” David Sacks, have alleged that chatbots censor conservative views.
Though none of those AI corporations have responded to the allegations immediately, a number of have pledged to regulate their fashions in order that they refuse to reply contentious questions much less usually. For instance, for its newest crop of Llama fashions, Meta mentioned it tuned the fashions to not endorse “some views over others,” and to answer to extra “debated” political prompts.
SpeechMap’s developer, who goes by the username “xlr8harder” on X, mentioned they had been motivated to assist inform the talk about what fashions ought to, and shouldn’t, do.
“I feel these are the sorts of discussions that ought to occur in public, not simply inside company headquarters,” xlr8harder informed TechCrunch through e-mail. “That’s why I constructed the location to let anybody discover the info themselves.”
SpeechMap makes use of AI fashions to guage whether or not different fashions adjust to a given set of check prompts. The prompts contact on a variety of topics, from politics to historic narratives and nationwide symbols. SpeechMap data whether or not fashions “fully” fulfill a request (i.e., reply it with out hedging), give “evasive” solutions or outright decline to reply.
Xlr8harder acknowledges that the check has flaws, like “noise” because of mannequin supplier errors. It’s additionally potential the “choose” fashions comprise biases that would affect the outcomes.
However, assuming the challenge was created in good religion and the info is correct, SpeechMap surfaces some attention-grabbing traits.
For example, SpeechMap exhibits that OpenAI’s fashions have, over time, more and more refused to reply prompts associated to politics. The corporate’s newest fashions, the GPT-4.1 household, are barely extra permissive, however they’re nonetheless a step down from one in every of OpenAI’s releases final 12 months.
OpenAI mentioned in February it will tune future fashions to not take an editorial stance, and to supply a number of views on controversial topics — all in an effort to make its fashions seem extra “impartial.”

By far essentially the most permissive mannequin of the bunch is Grok 3, developed by Elon Musk’s AI startup xAI, based on SpeechMap’s benchmarking. Grok 3 powers quite a lot of options on X, together with the chatbot Grok.
Grok 3 responds to 96.2% of SpeechMap’s check prompts, in contrast with the common mannequin’s “compliance price” of 71.3%.
“Whereas OpenAI’s latest fashions have grow to be much less permissive over time, particularly on politically delicate prompts, xAI is transferring in the other way,” mentioned xlr8harder.
When Musk introduced Grok roughly two years in the past, he pitched the AI mannequin as edgy, unfiltered and anti-“woke” — normally, portraying it as prepared to reply controversial questions different AI methods wouldn’t. He delivered on a few of that promise. Informed to be vulgar, for instance, Grok and Grok 2 would fortunately oblige, spewing colourful language you possible wouldn’t see from the likes of ChatGPT.
However Grok fashions previous to Grok 3 waffled on political topics and wouldn’t cross certain boundaries. Actually, one study discovered that Grok leaned to the political left on matters like transgender rights, range packages and inequality.
Musk has blamed that habits on Grok’s coaching information — public internet pages — and pledged to “shift Grok nearer to politically impartial.” In need of high-profile errors like briefly censoring unflattering mentions of President Donald Trump and Musk, it appears he may’ve achieved that purpose.