[2604.02338] LiME: Lightweight Mixture of Experts for Efficient Multimodal Multi-task Learning

[2604.02338] LiME: Lightweight Mixture of Experts for Efficient Multimodal Multi-task Learning

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2604.02338: LiME: Lightweight Mixture of Experts for Efficient Multimodal Multi-task Learning

Computer Science > Machine Learning arXiv:2604.02338 (cs) [Submitted on 1 Feb 2026] Title:LiME: Lightweight Mixture of Experts for Efficient Multimodal Multi-task Learning Authors:Md Kowsher, Haris Mansoor, Nusrat Jahan Prottasha, Ozlem Garibay, Victor Zhu, Zhengping Ji, Chen Chen View a PDF of the paper titled LiME: Lightweight Mixture of Experts for Efficient Multimodal Multi-task Learning, by Md Kowsher and 6 other authors View PDF Abstract:MoE-PEFT methods combine Mixture of Experts with parameter-efficient fine-tuning for multi-task adaptation, but require separate adapters per expert causing trainable parameters to scale linearly with expert count and limiting applicability to adapter-based architectures. We propose LiME (Lightweight Mixture of Experts), which achieves expert specialization through lightweight modulation rather than adapter replication. Instead of separate adapters, LiME uses a single shared PEFT module and modulates its output with lightweight expert vectors, reducing expert parameters while generalizing to any PEFT method. Notably, LiME introduces zero-parameter routing by leveraging existing frozen and adapted representations eliminating learned router parameters typically required per layer. Theoretically, we prove that (i) more experts preserve more task-relevant information and (ii) modulation approximates full expert-specific PEFT with bounded error. LiME further incorporates n-gram windowed routing and adaptive expert selection (Auto Top-K) b...

Originally published on April 06, 2026. Curated by AI News.

Related Articles

UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
University of Tartu thesis: transfer learning boosts Estonian AI models
Machine Learning

University of Tartu thesis: transfer learning boosts Estonian AI models

AI News - General · 4 min ·
ACM Prize in Computing Honors Matei Zaharia for Foundational Contributions to Data and Machine Learning Systems
Machine Learning

ACM Prize in Computing Honors Matei Zaharia for Foundational Contributions to Data and Machine Learning Systems

AI News - General · 6 min ·
Sam Altman's Coworkers Say He Can Barely Code and Misunderstands Basic Machine Learning Concepts
Machine Learning

Sam Altman's Coworkers Say He Can Barely Code and Misunderstands Basic Machine Learning Concepts

AI News - General · 2 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime