A new paper from AI lab Cohere, Stanford, MIT, and Ai2 accuses LM Enviornment, the group behind the favored crowdsourced AI benchmark Chatbot Enviornment, of serving to a choose group of AI firms obtain higher leaderboard scores on the expense of rivals.
Based on the authors, LM Enviornment allowed some industry-leading AI firms like Meta, OpenAI, Google, and Amazon to privately take a look at a number of variants of AI fashions, then not publish the scores of the bottom performers. This made it simpler for these firms to realize a prime spot on the platform’s leaderboard, although the chance was not afforded to each agency, the authors say.
“Solely a handful of [companies] had been advised that this non-public testing was obtainable, and the quantity of personal testing that some [companies] acquired is simply a lot greater than others,” stated Cohere’s VP of AI analysis and co-author of the research, Sara Hooker, in an interview with TechCrunch. “That is gamification.”
Created in 2023 as a tutorial analysis venture out of UC Berkeley, Chatbot Enviornment has change into a go-to benchmark for AI firms. It really works by placing solutions from two completely different AI fashions side-by-side in a “battle,” and asking customers to decide on the very best one. It’s not unusual to see unreleased fashions competing within the enviornment underneath a pseudonym.
Votes over time contribute to a mannequin’s rating — and, consequently, its placement on the Chatbot Enviornment leaderboard. Whereas many business actors take part in Chatbot Enviornment, LM Enviornment has lengthy maintained that its benchmark is an neutral and truthful one.
Nevertheless, that’s not what the paper’s authors say they uncovered.
One AI firm, Meta, was in a position to privately take a look at 27 mannequin variants on Chatbot Enviornment between January and March main as much as the tech big’s Llama 4 launch, the authors allege. At launch, Meta solely publicly revealed the rating of a single mannequin — a mannequin that occurred to rank close to the highest of the Chatbot Enviornment leaderboard.
Techcrunch occasion
Berkeley, CA
|
June 5
BOOK NOW

In an e-mail to TechCrunch, LM Enviornment Co-Founder and UC Berkeley Professor Ion Stoica stated that the research was filled with “inaccuracies” and “questionable evaluation.”
“We’re dedicated to truthful, community-driven evaluations, and invite all mannequin suppliers to submit extra fashions for testing and to enhance their efficiency on human choice,” stated LM Enviornment in an announcement supplied to TechCrunch. “If a mannequin supplier chooses to submit extra assessments than one other mannequin supplier, this doesn’t imply the second mannequin supplier is handled unfairly.”
Armand Joulin, a principal researcher at Google DeepMind, additionally famous in a post on X that among the research’s numbers had been inaccurate, claiming Google solely despatched one Gemma 3 AI mannequin to LM Enviornment for pre-release testing. Hooker responded to Joulin on X, promising the authors would make a correction.
Supposedly favored labs
The paper’s authors began conducting their analysis in November 2024 after studying that some AI firms had been probably being given preferential entry to Chatbot Enviornment. In complete, they measured greater than 2.8 million Chatbot Enviornment battles over a five-month stretch.
The authors say they discovered proof that LM Enviornment allowed sure AI firms, together with Meta, OpenAI, and Google, to gather extra knowledge from Chatbot Enviornment by having their fashions seem in the next variety of mannequin “battles.” This elevated sampling fee gave these firms an unfair benefit, the authors allege.
Utilizing extra knowledge from LM Enviornment may enhance a mannequin’s efficiency on Enviornment Laborious, one other benchmark LM Enviornment maintains, by 112%. Nevertheless, LM Enviornment stated in a post on X that Enviornment Laborious efficiency doesn’t immediately correlate to Chatbot Enviornment efficiency.
Hooker stated it’s unclear how sure AI firms may’ve acquired precedence entry, however that it’s incumbent on LM Enviornment to extend its transparency regardless.
In a post on X, LM Enviornment stated that a number of of the claims within the paper don’t replicate actuality. The group pointed to a blog post it revealed earlier this week indicating that fashions from non-major labs seem in additional Chatbot Enviornment battles than the research suggests.
One necessary limitation of the research is that it relied on “self-identification” to find out which AI fashions had been in non-public testing on Chatbot Enviornment. The authors prompted AI fashions a number of instances about their firm of origin, and relied on the fashions’ solutions to categorise them — a technique that isn’t foolproof.
Nevertheless, Hooker stated that when the authors reached out to LM Enviornment to share their preliminary findings, the group didn’t dispute them.
TechCrunch reached out to Meta, Google, OpenAI, and Amazon — all of which had been talked about within the research — for remark. None instantly responded.
LM Enviornment in sizzling water
Within the paper, the authors name on LM Enviornment to implement a variety of modifications aimed toward making Chatbot Enviornment extra “truthful.” For instance, the authors say, LM Enviornment may set a transparent and clear restrict on the variety of non-public assessments AI labs can conduct, and publicly disclose scores from these assessments.
In a post on X, LM Enviornment rejected these solutions, claiming it has revealed info on pre-release testing since March 2024. The benchmarking group additionally stated it “is mindless to point out scores for pre-release fashions which aren’t publicly obtainable,” as a result of the AI group can not take a look at the fashions for themselves.
The researchers additionally say LM Enviornment may modify Chatbot Enviornment’s sampling fee to make sure that all fashions within the enviornment seem in the identical variety of battles. LM Enviornment has been receptive to this advice publicly, and indicated that it’ll create a brand new sampling algorithm.
The paper comes weeks after Meta was caught gaming benchmarks in Chatbot Enviornment across the launch of its above-mentioned Llama 4 fashions. Meta optimized one of many Llama 4 fashions for “conversationality,” which helped it obtain a powerful rating on Chatbot Enviornment’s leaderboard. However the firm by no means launched the optimized mannequin — and the vanilla model ended up performing a lot worse on Chatbot Enviornment.
On the time, LM Enviornment stated Meta ought to have been extra clear in its strategy to benchmarking.
Earlier this month, LM Enviornment introduced it was launching a company, with plans to lift capital from traders. The research will increase scrutiny on non-public benchmark group’s — and whether or not they are often trusted to evaluate AI fashions with out company affect clouding the method.