[2602.13568] Who Do LLMs Trust? Human Experts Matter More Than Other LLMs

[2602.13568] Who Do LLMs Trust? Human Experts Matter More Than Other LLMs

arXiv - AI 3 min read Article

Summary

This paper explores how large language models (LLMs) prioritize feedback from human experts over other LLMs in decision-making tasks, revealing significant implications for AI trust and credibility.

Why It Matters

Understanding how LLMs interpret and prioritize information from different sources is crucial for developing more reliable AI systems. This research highlights the importance of human expertise in shaping LLM responses, which can impact applications in various fields such as education, healthcare, and automated decision-making.

Key Takeaways

  • LLMs show a preference for responses attributed to human experts over those from other LLMs.
  • Expert framing significantly influences LLM decision-making, even when the expert's input is incorrect.
  • The study indicates a form of credibility-sensitive social influence in LLMs across various decision domains.

Computer Science > Artificial Intelligence arXiv:2602.13568 (cs) [Submitted on 14 Feb 2026] Title:Who Do LLMs Trust? Human Experts Matter More Than Other LLMs Authors:Anooshka Bajaj, Zoran Tiganj View a PDF of the paper titled Who Do LLMs Trust? Human Experts Matter More Than Other LLMs, by Anooshka Bajaj and Zoran Tiganj View PDF HTML (experimental) Abstract:Large language models (LLMs) increasingly operate in environments where they encounter social information such as other agents' answers, tool outputs, or human recommendations. In humans, such inputs influence judgments in ways that depend on the source's credibility and the strength of consensus. This paper investigates whether LLMs exhibit analogous patterns of influence and whether they privilege feedback from humans over feedback from other LLMs. Across three binary decision-making tasks, reading comprehension, multi-step reasoning, and moral judgment, we present four instruction-tuned LLMs with prior responses attributed either to friends, to human experts, or to other LLMs. We manipulate whether the group is correct and vary the group size. In a second experiment, we introduce direct disagreement between a single human and a single LLM. Across tasks, models conform significantly more to responses labeled as coming from human experts, including when that signal is incorrect, and revise their answers toward experts more readily than toward other LLMs. These results reveal that expert framing acts as a strong prior...

Related Articles

Llms

Claude on Claude

The Story of Anthropic’s Latest Controversies Regarding the Business of Its Prized Creation… As Told by the Thing Itself. Editor’s note: ...

Reddit - Artificial Intelligence · 1 min ·
Llms

Cut Claude usage by ~85% in a job search pipeline (16k → 900 tokens/app) — here’s what worked

Like many here, I kept running into Claude usage limits when building anything non-trivial. I was working with a job search automation pi...

Reddit - Artificial Intelligence · 1 min ·
Llms

"Authoritarian Parents In Rationalist Clothes": a piece I wrote in December about alignment

Posted today in light of the Claude Mythos model card release. Originally I wrote this for r/ControlProblem but realized it was getting o...

Reddit - Artificial Intelligence · 1 min ·
Llms

AI joins the 8-hour work day as GLM ships 5.1 open source LLM, beating Opus 4.6 and GPT-5.4 on SWE-Bench Pro

AI Tools & Products ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime