[2602.18986] Quantifying Automation Risk in High-Automation AI Systems: A Bayesian Framework for Failure Propagation and Optimal Oversight

[2602.18986] Quantifying Automation Risk in High-Automation AI Systems: A Bayesian Framework for Failure Propagation and Optimal Oversight

arXiv - AI 4 min read Article

Summary

This paper presents a Bayesian framework for assessing automation risk in high-automation AI systems, focusing on failure propagation and optimal oversight to mitigate potential harms.

Why It Matters

As organizations increasingly rely on automated AI systems across various sectors, understanding and quantifying the risks associated with automation is crucial. This framework provides a structured approach to evaluate how failures can lead to significant harm, thus offering a pathway for better risk governance and oversight in AI deployments.

Key Takeaways

  • Introduces a Bayesian risk decomposition model for AI systems.
  • Isolates the conditional probability of failure propagation into harm.
  • Provides theoretical foundations for risk governance tools in AI.
  • Illustrates the framework with a case study of the Knight Capital incident.
  • Emphasizes the need for empirical validation across deployment domains.

Computer Science > Artificial Intelligence arXiv:2602.18986 (cs) [Submitted on 22 Feb 2026] Title:Quantifying Automation Risk in High-Automation AI Systems: A Bayesian Framework for Failure Propagation and Optimal Oversight Authors:Vishal Srivastava, Tanmay Sah View a PDF of the paper titled Quantifying Automation Risk in High-Automation AI Systems: A Bayesian Framework for Failure Propagation and Optimal Oversight, by Vishal Srivastava and 1 other authors View PDF HTML (experimental) Abstract:Organizations across finance, healthcare, transportation, content moderation, and critical infrastructure are rapidly deploying highly automated AI systems, yet they lack principled methods to quantify how increasing automation amplifies harm when failures occur. We propose a parsimonious Bayesian risk decomposition expressing expected loss as the product of three terms: the probability of system failure, the conditional probability that a failure propagates into harm given the automation level, and the expected severity of harm. This framework isolates a critical quantity -- the conditional probability that failures propagate into harm -- which captures execution and oversight risk rather than model accuracy alone. We develop complete theoretical foundations: formal proofs of the decomposition, a harm propagation equivalence theorem linking the harm propagation probability to observable execution controls, risk elasticity measures, efficient frontier analysis for automation policy, ...

Related Articles

Ai Safety

NHS staff resist using Palantir software. Staff reportedly cite ethics concerns, privacy worries, and doubt the platform adds much

submitted by /u/esporx [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

AI assistants are optimized to seem helpful. That is not the same thing as being helpful.

RLHF trains models on human feedback. Humans rate responses they like. And it turns out humans consistently rate confident, fluent, agree...

Reddit - Artificial Intelligence · 1 min ·
Computer Vision

House Democrat Questions Anthropic on AI Safety After Source Code Leak

Rep. Josh Gottheimer, who is generally tough on China, just sent a letter to Anthropic questioning their decision to reduce certain safet...

Reddit - Artificial Intelligence · 1 min ·
[2512.21106] Semantic Refinement with LLMs for Graph Representations
Llms

[2512.21106] Semantic Refinement with LLMs for Graph Representations

Abstract page for arXiv paper 2512.21106: Semantic Refinement with LLMs for Graph Representations

arXiv - Machine Learning · 4 min ·
More in Ai Safety: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime