Training-time intervention yields 63.4% blind-pair human preference at matched val-loss (1.2B params, 320 judgments, p = 1.98 × 10⁻⁵) [R]
About this article
TL;DR. I ran a blind A/B preference evaluation between two 1.2B-parameter LMs trained on identical data (same order, same seed, 30K steps / 3.9B tokens) - one with a Predictive-Coding-inspired precision-weighted gain function plus per-layer divergence-scaled gradients, one with standard cross-entropy. Smoothed val loss between the two is statistically indistinguishable (0.004-nat difference, well inside step-to-step noise). Ten judges (seven humans, three foundation models across Anthropic / ...
You've been blocked by network security.To continue, log in to your Reddit account or use your developer tokenIf you think you've been blocked by mistake, file a ticket below and we'll look into it.Log in File a ticket