[2602.21262] Under the Influence: Quantifying Persuasion and Vigilance in Large Language Models

[2602.21262] Under the Influence: Quantifying Persuasion and Vigilance in Large Language Models

arXiv - Machine Learning 4 min read Research

Summary

This paper investigates the interplay between persuasion and vigilance in Large Language Models (LLMs), revealing that these capacities are dissociable and critical for AI safety.

Why It Matters

As LLMs become integral in high-stakes decision-making, understanding their ability to persuade and discern information is vital. This research highlights the risks of misinformation and the need for monitoring LLMs' performance in both persuasion and vigilance to ensure safe AI deployment.

Key Takeaways

  • Persuasion and vigilance in LLMs are independent capacities.
  • High performance in tasks does not guarantee resistance to deception.
  • LLMs adjust token usage based on the nature of advice (benevolent vs. malicious).
  • Understanding these dynamics is crucial for future AI safety measures.
  • This study is the first to explore the relationship between persuasion, vigilance, and task performance in LLMs.

Computer Science > Computation and Language arXiv:2602.21262 (cs) [Submitted on 24 Feb 2026] Title:Under the Influence: Quantifying Persuasion and Vigilance in Large Language Models Authors:Sasha Robinson, Kerem Oktar, Katherine M. Collins, Ilia Sucholutsky, Kelsey R. Allen View a PDF of the paper titled Under the Influence: Quantifying Persuasion and Vigilance in Large Language Models, by Sasha Robinson and 4 other authors View PDF HTML (experimental) Abstract:With increasing integration of Large Language Models (LLMs) into areas of high-stakes human decision-making, it is important to understand the risks they introduce as advisors. To be useful advisors, LLMs must sift through large amounts of content, written with both benevolent and malicious intent, and then use this information to convince a user to take a specific action. This involves two social capacities: vigilance (the ability to determine which information to use, and which to discard) and persuasion (synthesizing the available evidence to make a convincing argument). While existing work has investigated these capacities in isolation, there has been little prior investigation of how these capacities may be linked. Here, we use a simple multi-turn puzzle-solving game, Sokoban, to study LLMs' abilities to persuade and be rationally vigilant towards other LLM agents. We find that puzzle-solving performance, persuasive capability, and vigilance are dissociable capacities in LLMs. Performing well on the game does n...

Related Articles

I Asked ChatGPT 500 Questions. Here Are the Ads I Saw Most Often | WIRED
Llms

I Asked ChatGPT 500 Questions. Here Are the Ads I Saw Most Often | WIRED

Ads are rolling out across the US on ChatGPT’s free tier. I asked OpenAI's bot 500 questions to see what these ads were like and how they...

Wired - AI · 9 min ·
Llms

Abacus.Ai Claw LLM consumes an incredible amount of credit without any usage :(

Three days ago, I clicked the "Deploy OpenClaw In Seconds" button to get an overview of the new service, but I didn't build any automatio...

Reddit - Artificial Intelligence · 1 min ·
Google’s Gemini AI app debuts in Hong Kong
Llms

Google’s Gemini AI app debuts in Hong Kong

Tech giant’s chatbot service tops Apple’s app store chart in the city.

AI Tools & Products · 2 min ·
Google Launches Gemini Import Tools to Poach Users From Rival AI Apps
Llms

Google Launches Gemini Import Tools to Poach Users From Rival AI Apps

Anyone looking to switch their AI assistant will find it surprisingly easy, as it only takes a few steps to move from A to B. This is not...

AI Tools & Products · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime