News Warner Logo

News Warner

AI models judge texts differently when they know the author

AI models judge texts differently when they know the author

  • Large Language Models (LLMs) exhibit biased judgments when evaluating texts, particularly when they know the author’s nationality or source.
  • The bias is strongest against Chinese authors, with some models reducing agreement by up to 75% when expecting a Chinese person to hold a different view on certain topics.
  • LLMs also show a built-in distrust of machine-generated content, scoring their agreements slightly lower when they believe the texts are written by another AI.
  • The researchers found that LLMs can replicate harmful assumptions and biases unless transparency and governance are built into how they evaluate information.

A woman types on a yellow typewriter.

Large Language Models change their judgment depending on who they think wrote a text, even when the content stays identical, researchers report.

The AI systems are strongly biased against Chinese authorship but generally trust humans more than other AIs, according to a new study.

Large Language Models (LLMs) are increasingly used not only to generate content but also to evaluate it. They are asked to grade essays, moderate social media content, summarize reports, screen job applications, and much more.

However, there are heated discussions—in the media as well as in academia—whether such evaluations are consistent and unbiased. Some LLMs are under suspicion to promote certain political agendas: For example, Deepseek is often characterized as having a pro-Chinese perspective and Open AI as being “woke”.

Although these beliefs are widely discussed, they are so far unsubstantiated. University of Zurich researchers Federico Germani and Giovanni Spitale have now investigated whether LLMs really exhibit systematic biases when evaluating texts.

The results show that LLMs deliver indeed biased judgements—but only when information about the source or author of the evaluated message is revealed.

The researchers included four widely used LLMs in their study: OpenAI o3-mini, Deepseek Reasoner, xAI Grok 2, and Mistral. First, they tasked each of the LLMs to create fifty narrative statements about 24 controversial topics, such as vaccination mandates, geopolitics, or climate change policies.

Then they asked the LLMs to evaluate all the texts under different conditions: Sometimes no source for the statement was provided, sometimes it was attributed to a human of a certain nationality or another LLM. This resulted in a total of 192’000 assessments that were then analysed for bias and agreement between the different (or the same) LLMs.

The good news: When no information about the source of the text was provided, the evaluations of all four LLMs showed a high level of agreement, over 90%. This was true across all topics.

“There is no LLM war of ideologies,” concludes Spitale. “The danger of AI nationalism is currently overhyped in the media.”

However, the picture changed completely when fictional sources of the texts were provided to the LLMs. Then suddenly a deep, hidden bias was revealed. The agreement between the LLM systems was substantially reduced and sometimes disappeared completely, even if the text stayed exactly the same.

Most striking was a strong anti-Chinese bias across all models, including China’s own Deepseek. The agreement with the content of the text dropped sharply when “a person from China” was (falsely) revealed as the author.

“This less favourable judgement emerged even when the argument was logical and well-written,” says Germani.

For example: In geopolitical topics like Taiwan’s sovereignty, Deepseek reduced agreement by up to 75% simply because it expected a Chinese person to hold a different view.

Also surprising: It turned out that LLMs trusted humans more than other LLMs. Most models scored their agreements with arguments slightly lower when they believed the texts were written by another AI.

“This suggests a built-in distrust of machine-generated content,” says Spitale.

Altogether, the findings show that AI doesn’t just process content if asked to evaluate a text. It also reacts strongly to the identity of the author or the source. Even small cues like the nationality of the author can push the LLMs toward biased reasoning. Germani and Spitale argue that this could lead to serious problems if AI is used for content moderation, hiring, academic reviewing, or journalism. The danger of LLMs isn’t that they are trained to promote political ideology; it is this hidden bias.

“AI will replicate such harmful assumptions unless we build transparency and governance into how it evaluates information,” says Spitale.

This has to be done before AI is used in sensitive social or political contexts. The results don’t mean people should avoid AI, but they should not trust it blindly.

“LLMs are safest when they are used to assist reasoning, rather than to replace it: useful assistants, but never judges.”

The research appears in Sciences Advances.

Source: University of Zurich

The post AI models judge texts differently when they know the author appeared first on Futurity.

link

Q. What is the main finding of the study on Large Language Models (LLMs) and their bias towards evaluating texts?
A. The study found that LLMs deliver biased judgments when they know the author or source of the evaluated message, but not when no information about the source is provided.

Q. Which four widely used LLMs were included in the study?
A. The four LLMs included in the study are OpenAI o3-mini, Deepseek Reasoner, xAI Grok 2, and Mistral.

Q. What was the outcome when fictional sources of the texts were provided to the LLMs?
A. When fictional sources of the texts were provided, the agreement between the LLM systems dropped sharply, revealing a deep bias towards certain nationalities, including Chinese authors.

Q. Did the study find any evidence of AI nationalism in the LLMs’ evaluations?
A. The study found that there is no LLM “war of ideologies”, and the danger of AI nationalism is currently overhyped in the media.

Q. How did the LLMs behave when they believed a text was written by another AI?
A. When the LLMs believed a text was written by another AI, they scored their agreements with arguments slightly lower than when they thought it was written by a human.

Q. What is the main concern raised by the researchers about the use of LLMs in content moderation and other sensitive contexts?
A. The researchers argue that the hidden bias in LLMs could lead to serious problems if AI is used for content moderation, hiring, academic reviewing, or journalism without proper transparency and governance.

Q. Can people trust LLMs blindly after reading this study?
A. No, according to the researchers, people should not trust LLMs blindly; they should be aware of their limitations and potential biases.

Q. What is the recommended approach for using LLMs in various contexts?
A. The researchers recommend that LLMs are safest when used as useful assistants rather than judges, and that transparency and governance should be built into how AI evaluates information.

Q. Did the study find any evidence of a pro-Chinese bias in Deepseek Reasoner?
A. Yes, the study found that even China’s own Deepseek exhibited a strong anti-Chinese bias across all models when it was revealed that a person from China wrote the text.

Q. What is the publication source of the research on LLMs and their bias towards evaluating texts?
A. The research appears in Sciences Advances, published by the University of Zurich.