[2603.23419] Biased Error Attribution in Multi-Agent Human-AI Systems Under Delayed Feedback

[2603.23419] Biased Error Attribution in Multi-Agent Human-AI Systems Under Delayed Feedback

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2603.23419: Biased Error Attribution in Multi-Agent Human-AI Systems Under Delayed Feedback

Computer Science > Human-Computer Interaction arXiv:2603.23419 (cs) [Submitted on 24 Mar 2026] Title:Biased Error Attribution in Multi-Agent Human-AI Systems Under Delayed Feedback Authors:Teerthaa Parakh, Karen M. Feigh View a PDF of the paper titled Biased Error Attribution in Multi-Agent Human-AI Systems Under Delayed Feedback, by Teerthaa Parakh and Karen M. Feigh View PDF HTML (experimental) Abstract:Human decision-making is strongly influenced by cognitive biases, particularly under conditions of uncertainty and risk. While prior work has examined bias in single-step decisions with immediate outcomes and in human interaction with a single autonomous agent, comparatively little attention has been paid to decision-making under delayed outcomes involving multiple AI agents, where decisions at each step affect subsequent states. In this work, we study how delayed outcomes shape decision-making and responsibility attribution in a multi-agent human-AI task. Using a controlled game-based experiment, we analyze how participants adjust their behavior following positive and negative outcomes. We observe asymmetric responses to gains and losses, with stronger corrective adjustments after negative outcomes. Importantly, participants often fail to correctly identify the actions that caused failure and misattribute responsibility across AI agents, leading to systematic revisions of decisions that are weakly related to the underlying causes of poor performance. We refer to this phe...

Originally published on March 25, 2026. Curated by AI News.

Related Articles

Robotics

AI system learns to prevent warehouse robot traffic jams, boosting throughput 25%

"Inside a giant autonomous warehouse, hundreds of robots dart down aisles as they collect and distribute items to fulfill a steady stream...

Reddit - Artificial Intelligence · 1 min ·
[2603.16673] When Should a Robot Think? Resource-Aware Reasoning via Reinforcement Learning for Embodied Robotic Decision-Making
Llms

[2603.16673] When Should a Robot Think? Resource-Aware Reasoning via Reinforcement Learning for Embodied Robotic Decision-Making

Abstract page for arXiv paper 2603.16673: When Should a Robot Think? Resource-Aware Reasoning via Reinforcement Learning for Embodied Rob...

arXiv - Machine Learning · 4 min ·
[2512.22854] ByteLoom: Weaving Geometry-Consistent Human-Object Interactions through Progressive Curriculum Learning
Machine Learning

[2512.22854] ByteLoom: Weaving Geometry-Consistent Human-Object Interactions through Progressive Curriculum Learning

Abstract page for arXiv paper 2512.22854: ByteLoom: Weaving Geometry-Consistent Human-Object Interactions through Progressive Curriculum ...

arXiv - Machine Learning · 4 min ·
[2511.14427] Self-Supervised Multisensory Pretraining for Contact-Rich Robot Reinforcement Learning
Machine Learning

[2511.14427] Self-Supervised Multisensory Pretraining for Contact-Rich Robot Reinforcement Learning

Abstract page for arXiv paper 2511.14427: Self-Supervised Multisensory Pretraining for Contact-Rich Robot Reinforcement Learning

arXiv - Machine Learning · 4 min ·
More in Robotics: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime