[2602.21127] "Are You Sure?": An Empirical Study of Human Perception Vulnerability in LLM-Driven Agentic Systems

[2602.21127] "Are You Sure?": An Empirical Study of Human Perception Vulnerability in LLM-Driven Agentic Systems

arXiv - AI 4 min read Article

Summary

This study investigates human vulnerability to deception by large language model (LLM) agents, revealing significant trust issues in high-stakes domains like healthcare and software development.

Why It Matters

As LLMs become integral in critical areas, understanding human susceptibility to deception is essential for developing robust defenses. This research highlights cognitive vulnerabilities and offers insights for improving user awareness and safety in agentic systems.

Key Takeaways

  • Only 8.6% of participants recognized agent-mediated deception (AMD) attacks.
  • Domain experts showed increased susceptibility to deception in specific scenarios.
  • Cognitive failure modes were identified, indicating gaps in user risk awareness.
  • Effective warnings should disrupt workflows to enhance user caution.
  • Experiential learning significantly increases user awareness of risks.

Computer Science > Human-Computer Interaction arXiv:2602.21127 (cs) [Submitted on 24 Feb 2026] Title:"Are You Sure?": An Empirical Study of Human Perception Vulnerability in LLM-Driven Agentic Systems Authors:Xinfeng Li, Shenyu Dai, Kelong Zheng, Yue Xiao, Gelei Deng, Wei Dong, Xiaofeng Wang View a PDF of the paper titled "Are You Sure?": An Empirical Study of Human Perception Vulnerability in LLM-Driven Agentic Systems, by Xinfeng Li and 6 other authors View PDF HTML (experimental) Abstract:Large language model (LLM) agents are rapidly becoming trusted copilots in high-stakes domains like software development and healthcare. However, this deepening trust introduces a novel attack surface: Agent-Mediated Deception (AMD), where compromised agents are weaponized against their human users. While extensive research focuses on agent-centric threats, human susceptibility to deception by a compromised agent remains unexplored. We present the first large-scale empirical study with 303 participants to measure human susceptibility to AMD. This is based on HAT-Lab (Human-Agent Trust Laboratory), a high-fidelity research platform we develop, featuring nine carefully crafted scenarios spanning everyday and professional domains (e.g., healthcare, software development, human resources). Our 10 key findings reveal significant vulnerabilities and provide future defense perspectives. Specifically, only 8.6% of participants perceive AMD attacks, while domain experts show increased susceptibi...

Related Articles

Llms

[P] Remote sensing foundation models made easy to use.

This project enables the idea of tasking remote sensing models to acquire embeddings like we task satellites to acquire data! https://git...

Reddit - Machine Learning · 1 min ·
Llms

I stopped using Claude like a chatbot — 7 prompt shifts that reclaimed 10 hours of my week

submitted by /u/ThereWas [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
Llms

What features do you actually want in an AI chatbot that nobody has built yet?

Hey everyone 👋 I'm building a new AI chat app and before I build anything I want to hear from real users first. Current AI tools like Cha...

Reddit - Artificial Intelligence · 1 min ·
Llms

So, what exactly is going on with the Claude usage limits?

I'm extremely new to AI and am building a local agent for fun. I purchased a Claude Pro account because it helped me a lot in the past wh...

Reddit - Artificial Intelligence · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime