LingVo.club
Level
LLMs change judgments when told who wrote a text — Level B2 — three white disc on brown surface

LLMs change judgments when told who wrote a textCEFR B2

25 Nov 2025

Level B2 – Upper-intermediate
5 min
272 words

Researchers at the University of Zurich analysed how author identity influences large language models’ evaluations of text. Federico Germani and Giovanni Spitale tested four LLMs—OpenAI o3-mini, Deepseek Reasoner, xAI Grok 2 and Mistral—by having each model generate fifty narrative statements on 24 controversial topics, from vaccination mandates to geopolitical and climate policy questions. The team then asked the models to rate the same statements under different source attributions and collected 192’000 assessments for analysis.

When no author information was provided, the models showed high mutual agreement—over 90%—prompting Spitale’s conclusion that “There is no LLM war of ideologies” and that media fears of “AI nationalism” may be overhyped. However, revealing a fictional author produced a deep hidden bias: agreement between systems fell sharply or even vanished, despite identical text. The most striking result was a strong anti-Chinese bias across all models, including China’s own Deepseek; agreement with content dropped when “a person from China” was given as author. In some geopolitical questions, for example Taiwan’s sovereignty, Deepseek reduced agreement by up to 75% because it expected a different view from a Chinese author.

The study also found that most models gave slightly lower agreement scores when they believed a text was written by another AI, indicating a built-in distrust of machine-generated content. Germani and Spitale warn that such hidden biases matter for content moderation, hiring, academic review and journalism. They call for transparency and governance in AI evaluation and recommend using LLMs to assist reasoning—not to replace human judgment, saying they can be “useful assistants, but never judges.” The research appears in Science Advances.

Difficult words

  • evaluationjudgment or measurement of quality or value
    evaluations
  • biasunfair preference or judgment for or against someone
  • attributionstatement assigning a text to a specific author
    attributions
  • agreementmeasure of how similar opinions or ratings are
  • distrustlack of trust or confidence in someone or something
  • content moderationprocess of reviewing and removing online content
  • transparencyopen sharing of information and reasons
  • governancerules and processes that control a system

Tip: hover, focus or tap highlighted words in the article to see quick definitions while you read or listen.

Discussion questions

  • How could hidden biases in LLM evaluations affect content moderation or hiring decisions? Give examples.
  • What steps might improve transparency and governance in AI evaluation in your opinion?
  • Do you agree that LLMs should assist reasoning but not replace human judgment? Why or why not?

Related articles