[2510.23557] Minimizing Human Intervention in Online Classification

[2510.23557] Minimizing Human Intervention in Online Classification

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2510.23557: Minimizing Human Intervention in Online Classification

Statistics > Machine Learning arXiv:2510.23557 (stat) [Submitted on 27 Oct 2025 (v1), last revised 1 May 2026 (this version, v2)] Title:Minimizing Human Intervention in Online Classification Authors:William Réveillard, Vasileios Saketos, Alexandre Proutiere, Richard Combes View a PDF of the paper titled Minimizing Human Intervention in Online Classification, by William R\'eveillard and 3 other authors View PDF HTML (experimental) Abstract:Training or fine-tuning large language model (LLM)-based systems often requires costly human feedback, yet there is limited understanding of how to minimize such intervention while maintaining strong error guarantees. We study this problem for LLM-based classification systems in an active learning framework: an agent sequentially labels $d$-dimensional query embeddings drawn i.i.d. from an unknown distribution by either calling a costly expert or guessing with no feedback, with the goal of minimizing regret relative to an oracle with free expert access. When the horizon $T$ is at least exponential in the embedding dimension $d$, the geometry of the class regions can be learned. In this regime, we propose the Conservative Hull-based Classifier (CHC), which maintains convex hulls of expert-labeled queries and calls the expert when a query lands outside all known hulls. CHC attains $\mathcal{O}(\log^d T)$ regret in $T$ and is minimax optimal for $d=1$. Otherwise, the geometry cannot be reliably learned in general. We show that for queries dr...

Originally published on May 04, 2026. Curated by AI News.

Related Articles

Llms

The recursive self, explained

looking for anyone to give any critiques or tell me that something here is incorrect. this is the work of a year how I scaffold on a true...

Reddit - Artificial Intelligence · 1 min ·
Llms

Excellent discussion about LLM scaling [D]

I came across an excellent in depth discussion of memory and compute scaling analysis for LLMs. One takeaway is that running LLMs locally...

Reddit - Machine Learning · 1 min ·
[2602.03216] Token Sparse Attention: Efficient Long-Context Inference with Interleaved Token Selection
Llms

[2602.03216] Token Sparse Attention: Efficient Long-Context Inference with Interleaved Token Selection

Abstract page for arXiv paper 2602.03216: Token Sparse Attention: Efficient Long-Context Inference with Interleaved Token Selection

arXiv - Machine Learning · 4 min ·
[2601.21214] Scaling Reasoning Hop Exposes Weaknesses: Demystifying and Improving Hop Generalization in Large Language Models
Llms

[2601.21214] Scaling Reasoning Hop Exposes Weaknesses: Demystifying and Improving Hop Generalization in Large Language Models

Abstract page for arXiv paper 2601.21214: Scaling Reasoning Hop Exposes Weaknesses: Demystifying and Improving Hop Generalization in Larg...

arXiv - Machine Learning · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime