[2509.02452] Do LLMs Adhere to Label Definitions? Examining Their Receptivity to External Label Definitions

[2509.02452] Do LLMs Adhere to Label Definitions? Examining Their Receptivity to External Label Definitions

arXiv - Machine Learning 3 min read Article

Summary

This article investigates whether large language models (LLMs) adhere to external label definitions or rely on internal representations, revealing mixed results across various tasks.

Why It Matters

Understanding how LLMs process external definitions is crucial for improving their accuracy and explainability. This research highlights the inconsistency in LLMs' integration of external knowledge, which has implications for their application in both general and domain-specific tasks.

Key Takeaways

  • LLMs often default to internal representations over external definitions.
  • Explicit label definitions can enhance accuracy, especially in domain-specific tasks.
  • The integration of external definitions into LLMs is not guaranteed or consistent.
  • Controlled experiments reveal varying receptivity to label definitions.
  • A deeper understanding of LLMs' processing of external knowledge is needed.

Computer Science > Computation and Language arXiv:2509.02452 (cs) [Submitted on 2 Sep 2025 (v1), last revised 24 Feb 2026 (this version, v3)] Title:Do LLMs Adhere to Label Definitions? Examining Their Receptivity to External Label Definitions Authors:Seyedali Mohammadi, Bhaskara Hanuma Vedula, Hemank Lamba, Edward Raff, Ponnurangam Kumaraguru, Francis Ferraro, Manas Gaur View a PDF of the paper titled Do LLMs Adhere to Label Definitions? Examining Their Receptivity to External Label Definitions, by Seyedali Mohammadi and 6 other authors View PDF HTML (experimental) Abstract:Do LLMs genuinely incorporate external definitions, or do they primarily rely on their parametric knowledge? To address these questions, we conduct controlled experiments across multiple explanation benchmark datasets (general and domain-specific) and label definition conditions, including expert-curated, LLM-generated, perturbed, and swapped definitions. Our results reveal that while explicit label definitions can enhance accuracy and explainability, their integration into an LLM's task-solving processes is neither guaranteed nor consistent, suggesting reliance on internalized representations in many cases. Models often default to their internal representations, particularly in general tasks, whereas domain-specific tasks benefit more from explicit definitions. These findings underscore the need for a deeper understanding of how LLMs process external knowledge alongside their pre-existing capabilities....

Related Articles

Llms

[R] Is autoresearch really better than classic hyperparameter tuning?

We did experiments comparing Optuna & autoresearch. Autoresearch converges faster, is more cost-efficient, and even generalizes bette...

Reddit - Machine Learning · 1 min ·
Llms

Claude Source Code?

Has anyone been able to successfully download the leaked source code yet? I've not been able to find it. If anyone has, please reach out....

Reddit - Artificial Intelligence · 1 min ·
Llms

[R] Solving the Jane Street Dormant LLM Challenge: A Systematic Approach to Backdoor Discovery

Submitted by: Adam Kruger Date: March 23, 2026 Models Solved: 3/3 (M1, M2, M3) + Warmup Background When we first encountered the Jane Str...

Reddit - Machine Learning · 1 min ·
Llms

[P] Gemma 4 running on NVIDIA B200 and AMD MI355X from the same inference stack, 15% throughput gain over vLLM on Blackwell

Google DeepMind dropped Gemma 4 today: Gemma 4 31B: dense, 256K context, redesigned architecture targeting efficiency and long-context qu...

Reddit - Machine Learning · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime