[2602.13934] Why Code, Why Now: Learnability, Computability, and the Real Limits of Machine Learning
Summary
The paper discusses the learnability and computability limits of machine learning, emphasizing the structured feedback of code generation compared to reinforcement learning, and proposes a hierarchy of learnability.
Why It Matters
Understanding the limits of machine learning is crucial for researchers and practitioners. This paper challenges the assumption that simply scaling models will address all challenges, providing a framework for evaluating the learnability of tasks, which can guide future research and applications in AI.
Key Takeaways
- Code generation benefits from structured feedback, unlike reinforcement learning.
- A five-level hierarchy of learnability is proposed based on information structure.
- The ceiling on ML progress is more about task learnability than model size.
- Supervised learning on code scales predictably, while reinforcement learning does not.
- The assumption that scaling alone will solve ML challenges should be scrutinized.
Computer Science > Machine Learning arXiv:2602.13934 (cs) [Submitted on 15 Feb 2026] Title:Why Code, Why Now: Learnability, Computability, and the Real Limits of Machine Learning Authors:Zhimin Zhao View a PDF of the paper titled Why Code, Why Now: Learnability, Computability, and the Real Limits of Machine Learning, by Zhimin Zhao View PDF HTML (experimental) Abstract:Code generation has progressed more reliably than reinforcement learning, largely because code has an information structure that makes it learnable. Code provides dense, local, verifiable feedback at every token, whereas most reinforcement learning problems do not. This difference in feedback quality is not binary but graded. We propose a five-level hierarchy of learnability based on information structure and argue that the ceiling on ML progress depends less on model size than on whether a task is learnable at all. The hierarchy rests on a formal distinction among three properties of computational problems (expressibility, computability, and learnability). We establish their pairwise relationships, including where implications hold and where they fail, and present a unified template that makes the structural differences explicit. The analysis suggests why supervised learning on code scales predictably while reinforcement learning does not, and why the common assumption that scaling alone will solve remaining ML challenges warrants scrutiny. Subjects: Machine Learning (cs.LG); Computation and Language (cs.CL...