[2601.21439] The Paradox of Robustness: Decoupling Rule-Based Logic from Affective Noise in High-Stakes Decision-Making

[2601.21439] The Paradox of Robustness: Decoupling Rule-Based Logic from Affective Noise in High-Stakes Decision-Making

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2601.21439: The Paradox of Robustness: Decoupling Rule-Based Logic from Affective Noise in High-Stakes Decision-Making

Computer Science > Artificial Intelligence arXiv:2601.21439 (cs) [Submitted on 29 Jan 2026 (v1), last revised 4 Apr 2026 (this version, v2)] Title:The Paradox of Robustness: Decoupling Rule-Based Logic from Affective Noise in High-Stakes Decision-Making Authors:Jon Chun, Katherine Elkins View a PDF of the paper titled The Paradox of Robustness: Decoupling Rule-Based Logic from Affective Noise in High-Stakes Decision-Making, by Jon Chun and 1 other authors View PDF HTML (experimental) Abstract:While Large Language Models (LLMs) are widely documented to be sensitive to minor prompt perturbations and prone to sycophantic alignment, their robustness in consequential, rule-bound decision-making remains under-explored. We uncover a striking "Paradox of Robustness": despite their known lexical brittleness, aligned LLMs exhibit strong robustness to emotional framing effects in rule-bound institutional decision-making. Using a controlled perturbation framework across three high-stakes domains (healthcare, finance, and education), we find a negligible effect size (Cohen's h = 0.003) compared to the substantial biases observed in analogous human contexts (h in [0.3, 0.8]), approximately two orders of magnitude smaller. This invariance persists across eight models with diverse training paradigms, suggesting the mechanisms driving sycophancy and prompt sensitivity do not translate to failures in logical constraint satisfaction. While LLMs may be "brittle" to how a query is formatted, t...

Originally published on April 07, 2026. Curated by AI News.

Related Articles

ChatGPT downloads are slowing — and may cause problems for OpenAI’s IPO | The Verge
Llms

ChatGPT downloads are slowing — and may cause problems for OpenAI’s IPO | The Verge

Data from Sensor Tower shows ChatGPT’s growth is slowing down, as Claude and other competitors’ growth is increasing, just as OpenAI is p...

The Verge - AI · 4 min ·
Larry Ellison’s betting everything on OpenAI. Will it pay off or pop the bubble? | The Verge
Llms

Larry Ellison’s betting everything on OpenAI. Will it pay off or pop the bubble? | The Verge

Larry Ellison and Oracle have staked their future on a data center deal with OpenAI and a big bet that enterprise AI will pay off.

The Verge - AI · 32 min ·
Llms

Google just released Deep Research Max — an autonomous research agent that writes expert-grade reports on its own

Google quietly dropped something interesting last week. They updated their Deep Research agent (available via Gemini API) and introduced ...

Reddit - Artificial Intelligence · 1 min ·
When Robots Have Their ChatGPT Moment, Remember These Pincers | WIRED
Llms

When Robots Have Their ChatGPT Moment, Remember These Pincers | WIRED

From sorting chicken nuggets to screwing in light bulbs, Eka’s robots are eerily lifelike. But do they have real physical smarts?

Wired - AI · 13 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime