[2603.01526] Scalable Multi-Task Low-Rank Model Adaptation

[2603.01526] Scalable Multi-Task Low-Rank Model Adaptation

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2603.01526: Scalable Multi-Task Low-Rank Model Adaptation

Computer Science > Machine Learning arXiv:2603.01526 (cs) [Submitted on 2 Mar 2026] Title:Scalable Multi-Task Low-Rank Model Adaptation Authors:Zichen Tian, Antoine Ledent, Qianru Sun View a PDF of the paper titled Scalable Multi-Task Low-Rank Model Adaptation, by Zichen Tian and 2 other authors View PDF HTML (experimental) Abstract:Scaling multi-task low-rank adaptation (LoRA) to a large number of tasks induces catastrophic performance degradation, such as an accuracy drop from 88.2% to 2.0% on DOTA when scaling from 5 to 15 tasks. This failure is due to parameter and representation misalignment. We find that existing solutions, like regularization and dynamic routing, fail at scale because they are constrained by a fundamental trade-off: strengthening regularization to reduce inter-task conflict inadvertently suppresses the essential feature discrimination required for effective routing. In this work, we identify two root causes for this trade-off. First, uniform regularization disrupts inter-task knowledge sharing: shared underlying knowledge concentrates in high-SV components (89% alignment on Flanv2->BBH). Uniform regularization forces high-SV components to update in orthogonal directions, directly disrupting the shared knowledge. Second, Conflict Amplification: Applying LoRA at the component-level (e.g., W_q, W_v) amplifies gradient conflicts; we show block-level adaptation reduces this conflict by 76% with only 50% parameters. Based on these insights, we propose mtL...

Originally published on March 03, 2026. Curated by AI News.

Related Articles

Machine Learning

Making an AI native sovereign computational stack

I’ve been working on a personal project that ended up becoming a kind of full computing stack: identity / trust protocol decentralized ch...

Reddit - Artificial Intelligence · 1 min ·
Llms

An attack class that passes every current LLM filter - no payload, no injection signature, no log trace

https://shapingrooms.com/research I published a paper today on something I've been calling postural manipulation. The short version: ordi...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

What tools are sr MLEs using? (clawdbot, openspec, wispr) [D]

I'm already blasting cursor, but I want to level up my output. I heard that these kind of AI tools and workflows are being asked in SF. W...

Reddit - Machine Learning · 1 min ·
Machine Learning

[R] looking for academic collaborators

hey there, i am currently working with a research group at auckland university. we are currently working on neurodegenerative diseases - ...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime