[2602.21919] Learning in the Null Space: Small Singular Values for Continual Learning

[2602.21919] Learning in the Null Space: Small Singular Values for Continual Learning

arXiv - Machine Learning 4 min read Article

Summary

The paper presents NESS, a novel continual learning method that leverages small singular values to maintain orthogonality in weight updates, addressing catastrophic forgetting in machine learning.

Why It Matters

Continual learning is crucial for developing AI systems that can adapt over time without losing previously acquired knowledge. This research provides a new approach that enhances learning efficiency and stability, which is vital for real-world applications in AI.

Key Takeaways

  • NESS utilizes small singular values to construct a null space for weight updates.
  • The method reduces catastrophic forgetting while enabling adaptation to new tasks.
  • Theoretical analysis and experiments show competitive performance across benchmark datasets.
  • The approach maintains stability in accuracy, crucial for continual learning applications.
  • A single trainable matrix per task simplifies the learning process.

Computer Science > Machine Learning arXiv:2602.21919 (cs) [Submitted on 25 Feb 2026] Title:Learning in the Null Space: Small Singular Values for Continual Learning Authors:Cuong Anh Pham, Praneeth Vepakomma, Samuel Horváth View a PDF of the paper titled Learning in the Null Space: Small Singular Values for Continual Learning, by Cuong Anh Pham and 2 other authors View PDF HTML (experimental) Abstract:Alleviating catastrophic forgetting while enabling further learning is a primary challenge in continual learning (CL). Orthogonal-based training methods have gained attention for their efficiency and strong theoretical properties, and many existing approaches enforce orthogonality through gradient projection. In this paper, we revisit orthogonality and exploit the fact that small singular values correspond to directions that are nearly orthogonal to the input space of previous tasks. Building on this principle, we introduce NESS (Null-space Estimated from Small Singular values), a CL method that applies orthogonality directly in the weight space rather than through gradient manipulation. Specifically, NESS constructs an approximate null space using the smallest singular values of each layer's input representation and parameterizes task-specific updates via a compact low-rank adaptation (LoRA-style) formulation constrained to this subspace. The subspace basis is fixed to preserve the null-space constraint, and only a single trainable matrix is learned for each task. This design...

Related Articles

Llms

Is the Mirage Effect a bug, or is it Geometric Reconstruction in action? A framework for why VLMs perform better "hallucinating" than guessing, and what that may tell us about what's really inside these models

Last week, a team from Stanford and UCSF (Asadi, O'Sullivan, Fei-Fei Li, Euan Ashley et al.) dropped two companion papers. The first, MAR...

Reddit - Artificial Intelligence · 1 min ·
Yupp shuts down after raising $33M from a16z crypto's Chris Dixon | TechCrunch
Machine Learning

Yupp shuts down after raising $33M from a16z crypto's Chris Dixon | TechCrunch

Less than a year after launching, with checks from some of the biggest names in Silicon Valley, crowdsourced AI model feedback startup Yu...

TechCrunch - AI · 4 min ·
Machine Learning

[R] Fine-tuning services report

If you have some data and want to train or run a small custom model but don't have powerful enough hardware for training, fine-tuning ser...

Reddit - Machine Learning · 1 min ·
Machine Learning

[D] Does ML have a "bible"/reference textbook at the Intermediate/Advanced level?

Hello, everyone! This is my first time posting here and I apologise if the question is, perhaps, a bit too basic for this sub-reddit. A b...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime