[2603.07926] IMSE: Intrinsic Mixture of Spectral Experts Fine-tuning for Test-Time Adaptation

[2603.07926] IMSE: Intrinsic Mixture of Spectral Experts Fine-tuning for Test-Time Adaptation

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2603.07926: IMSE: Intrinsic Mixture of Spectral Experts Fine-tuning for Test-Time Adaptation

Computer Science > Computer Vision and Pattern Recognition arXiv:2603.07926 (cs) [Submitted on 9 Mar 2026 (v1), last revised 14 Apr 2026 (this version, v3)] Title:IMSE: Intrinsic Mixture of Spectral Experts Fine-tuning for Test-Time Adaptation Authors:Sunghyun Baek, Jaemyung Yu, Seunghee Koh, Minsu Kim, Hyeonseong Jeon, Junmo Kim View a PDF of the paper titled IMSE: Intrinsic Mixture of Spectral Experts Fine-tuning for Test-Time Adaptation, by Sunghyun Baek and 5 other authors View PDF HTML (experimental) Abstract:Test-time adaptation (TTA) has been widely explored to prevent performance degradation when test data differ from the training distribution. However, fully leveraging the rich representations of large pretrained models with minimal parameter updates remains underexplored. In this paper, we propose Intrinsic Mixture of Spectral Experts (IMSE) that leverages the spectral experts inherently embedded in Vision Transformers. We decompose each linear layer via singular value decomposition (SVD) and adapt only the singular values, while keeping the singular vectors fixed. We further identify a key limitation of entropy minimization in TTA: it often induces feature collapse, causing the model to rely on domain-specific features rather than class-discriminative features. To address this, we propose a diversity maximization loss based on expert-input alignment, which encourages diverse utilization of spectral experts during adaptation. In the continual test-time adaptation...

Originally published on April 15, 2026. Curated by AI News.

Related Articles

Machine Learning

Trained a Qwen2.5-0.5B-Instruct bf16 model on Reddit post summarization task with GRPO written from scratch in PyTorch - updates! [P]

So, yesterday run was a success and I did get an avg rollout length of about 64 tokens as attached in the image! This was with quality_re...

Reddit - Machine Learning · 1 min ·
Machine Learning

Thoughts and experience on ML journals [D]

Recently I’ve been thinking about shifting from conferences to journals due to a few bad experiences with ML conferences reviewing proces...

Reddit - Machine Learning · 1 min ·
Llms

Gemini Robotics-ER 1.6: Powering real-world robotics tasks through enhanced embodied reasoning

Gemini Robotics-ER 1.6 is a significant upgrade to the reasoning-first model that enables robots to understand their environments with un...

Reddit - Artificial Intelligence · 1 min ·
[2603.10652] Are Video Reasoning Models Ready to Go Outside?
Llms

[2603.10652] Are Video Reasoning Models Ready to Go Outside?

Abstract page for arXiv paper 2603.10652: Are Video Reasoning Models Ready to Go Outside?

arXiv - AI · 4 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime