[2602.22070] Language Models Exhibit Inconsistent Biases Towards Algorithmic Agents and Human Experts

[2602.22070] Language Models Exhibit Inconsistent Biases Towards Algorithmic Agents and Human Experts

arXiv - AI 4 min read Research

Summary

This study explores how large language models (LLMs) exhibit inconsistent biases towards algorithmic agents and human experts in decision-making tasks, revealing significant implications for AI deployment.

Why It Matters

Understanding the biases of LLMs is crucial as they are increasingly used in decision-making roles. This research highlights the potential risks of deploying AI systems that may favor human experts over algorithms or vice versa, impacting trust and effectiveness in high-stakes scenarios.

Key Takeaways

  • LLMs show higher trust ratings for human experts compared to algorithms.
  • In incentivized betting scenarios, LLMs may favor algorithms even when they perform worse.
  • The study underscores the importance of task presentation formats in evaluating AI biases.
  • Inconsistent biases in LLMs can affect decision-making in critical applications.
  • Robust scrutiny of LLM evaluation methods is essential for AI safety.

Computer Science > Artificial Intelligence arXiv:2602.22070 (cs) [Submitted on 25 Feb 2026] Title:Language Models Exhibit Inconsistent Biases Towards Algorithmic Agents and Human Experts Authors:Jessica Y. Bo, Lillio Mok, Ashton Anderson View a PDF of the paper titled Language Models Exhibit Inconsistent Biases Towards Algorithmic Agents and Human Experts, by Jessica Y. Bo and 2 other authors View PDF HTML (experimental) Abstract:Large language models are increasingly used in decision-making tasks that require them to process information from a variety of sources, including both human experts and other algorithmic agents. How do LLMs weigh the information provided by these different sources? We consider the well-studied phenomenon of algorithm aversion, in which human decision-makers exhibit bias against predictions from algorithms. Drawing upon experimental paradigms from behavioural economics, we evaluate how eightdifferent LLMs delegate decision-making tasks when the delegatee is framed as a human expert or an algorithmic agent. To be inclusive of different evaluation formats, we conduct our study with two task presentations: stated preferences, modeled through direct queries about trust towards either agent, and revealed preferences, modeled through providing in-context examples of the performance of both agents. When prompted to rate the trustworthiness of human experts and algorithms across diverse tasks, LLMs give higher ratings to the human expert, which correlates...

Related Articles

I Asked ChatGPT 500 Questions. Here Are the Ads I Saw Most Often | WIRED
Llms

I Asked ChatGPT 500 Questions. Here Are the Ads I Saw Most Often | WIRED

Ads are rolling out across the US on ChatGPT’s free tier. I asked OpenAI's bot 500 questions to see what these ads were like and how they...

Wired - AI · 9 min ·
Llms

Abacus.Ai Claw LLM consumes an incredible amount of credit without any usage :(

Three days ago, I clicked the "Deploy OpenClaw In Seconds" button to get an overview of the new service, but I didn't build any automatio...

Reddit - Artificial Intelligence · 1 min ·
Google’s Gemini AI app debuts in Hong Kong
Llms

Google’s Gemini AI app debuts in Hong Kong

Tech giant’s chatbot service tops Apple’s app store chart in the city.

AI Tools & Products · 2 min ·
Google Launches Gemini Import Tools to Poach Users From Rival AI Apps
Llms

Google Launches Gemini Import Tools to Poach Users From Rival AI Apps

Anyone looking to switch their AI assistant will find it surprisingly easy, as it only takes a few steps to move from A to B. This is not...

AI Tools & Products · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime