[2411.00623] Replay-Free Continual Low-Rank Adaptation with Dynamic Memory

[2411.00623] Replay-Free Continual Low-Rank Adaptation with Dynamic Memory

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2411.00623: Replay-Free Continual Low-Rank Adaptation with Dynamic Memory

Computer Science > Computer Vision and Pattern Recognition arXiv:2411.00623 (cs) [Submitted on 1 Nov 2024 (v1), last revised 24 Mar 2026 (this version, v4)] Title:Replay-Free Continual Low-Rank Adaptation with Dynamic Memory Authors:Huancheng Chen, Jingtao Li, Weiming Zhuang, Chen Chen, Lingjuan Lyu View a PDF of the paper titled Replay-Free Continual Low-Rank Adaptation with Dynamic Memory, by Huancheng Chen and Jingtao Li and Weiming Zhuang and Chen Chen and Lingjuan Lyu View PDF HTML (experimental) Abstract:We revisit continual learning~(CL), which enables pre-trained vision transformers (ViTs) to sequentially fine-tune on new downstream tasks over time. However, as the scale of these models increases, catastrophic forgetting remains a more serious challenge. Recent studies highlight a crossover between CL techniques and parameter-efficient fine-tuning (PEFT), which focuses on fine-tuning only a small set of trainable parameters to adapt to downstream tasks, such as low-rank adaptation (LoRA). While LoRA achieves faster convergence and requires fewer trainable parameters, it has seldom been explored in the context of continual learning. To address this gap, we propose a novel PEFT-CL method called Dual Low-Rank Adaptation (DualLoRA), which introduces both an orthogonal LoRA adapter and a residual LoRA adapter parallel to pre-trained weights in each layer. These components are orchestrated by a dynamic memory mechanism to strike a balance between stability and plasticity...

Originally published on March 25, 2026. Curated by AI News.

Related Articles

Machine Learning

[P] Fused MoE Dispatch in Pure Triton: Beating CUDA-Optimized Megablocks at Inference Batch Sizes

I built a fused MoE dispatch kernel in pure Triton that handles the full forward pass for Mixture-of-Experts models. No CUDA, no vendor-s...

Reddit - Machine Learning · 1 min ·
Machine Learning

[D] ICML Rebuttal Question

I am currently working on my response on the rebuttal acknowledgments for ICML and I doubting how to handle the strawman argument of that...

Reddit - Machine Learning · 1 min ·
Machine Learning

[D] ML researcher looking to switch to a product company.

Hey, I am an AI researcher currently working in a deep tech company as a data scientist. Prior to this, I was doing my PhD. My current ro...

Reddit - Machine Learning · 1 min ·
Machine Learning

Building behavioural response models of public figures using Brain scan data (Predict their next move using psychological modelling) [P]

Hey guys, I’m the same creator of Netryx V2, the geolocation tool. I’ve been working on something new called COGNEX. It learns how a pers...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime