[2506.02630] Hyperbolic Aware Minimization: Implicit Bias for Sparsity
About this article
Abstract page for arXiv paper 2506.02630: Hyperbolic Aware Minimization: Implicit Bias for Sparsity
Computer Science > Machine Learning arXiv:2506.02630 (cs) [Submitted on 3 Jun 2025 (v1), last revised 2 Mar 2026 (this version, v2)] Title:Hyperbolic Aware Minimization: Implicit Bias for Sparsity Authors:Tom Jacobs, Advait Gadhikar, Celia Rubio-Madrigal, Rebekka Burkholz View a PDF of the paper titled Hyperbolic Aware Minimization: Implicit Bias for Sparsity, by Tom Jacobs and 3 other authors View PDF Abstract:Understanding the implicit bias of optimization algorithms is key to explaining and improving the generalization of deep models. The hyperbolic implicit bias induced by pointwise overparameterization promotes sparsity, but also yields a small inverse Riemannian metric near zero, slowing down parameter movement and impeding meaningful parameter sign flips. To overcome this obstacle, we propose Hyperbolic Aware Minimization (HAM), which alternates a standard optimizer step with a lightweight hyperbolic mirror step. The mirror step incurs less compute and memory than pointwise overparameterization, reproduces its beneficial hyperbolic geometry for feature learning, and mitigates the small-inverse-metric bottleneck. Our characterization of the implicit bias in the context of underdetermined linear regression provides insights into the mechanism how HAM consistently increases performance --even in the case of dense training, as we demonstrate in experiments with standard vision benchmarks. HAM is especially effective in combination with different sparsification methods, ...