[2603.01759] Meta-Learning Hyperparameters for Parameter Efficient Fine-Tuning

[2603.01759] Meta-Learning Hyperparameters for Parameter Efficient Fine-Tuning

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2603.01759: Meta-Learning Hyperparameters for Parameter Efficient Fine-Tuning

Computer Science > Machine Learning arXiv:2603.01759 (cs) [Submitted on 2 Mar 2026] Title:Meta-Learning Hyperparameters for Parameter Efficient Fine-Tuning Authors:Zichen Tian, Yaoyao Liu, Qianru Sun View a PDF of the paper titled Meta-Learning Hyperparameters for Parameter Efficient Fine-Tuning, by Zichen Tian and 2 other authors View PDF HTML (experimental) Abstract:Training large foundation models from scratch for domain-specific applications is almost impossible due to data limits and long-tailed distributions -- taking remote sensing (RS) as an example. Fine-tuning natural image pre-trained models on RS images is a straightforward solution. To reduce computational costs and improve performance on tail classes, existing methods apply parameter-efficient fine-tuning (PEFT) techniques, such as LoRA and AdaptFormer. However, we observe that fixed hyperparameters -- such as intra-layer positions, layer depth, and scaling factors, can considerably hinder PEFT performance, as fine-tuning on RS images proves highly sensitive to these settings. To address this, we propose MetaPEFT, a method incorporating adaptive scalers that dynamically adjust module influence during fine-tuning. MetaPEFT dynamically adjusts three key factors of PEFT on RS images: module insertion, layer selection, and module-wise learning rates, which collectively control the influence of PEFT modules across the network. We conduct extensive experiments on three transfer-learning scenarios and five datasets ...

Originally published on March 03, 2026. Curated by AI News.

Related Articles

Llms

We benchmarked TranslateGemma against 5 other LLMs on subtitle translation across 6 languages. At first glance the numbers told a clean story, but then human QA added a chapter. [D]

We evaluated six models on English subtitle translation into Spanish, Japanese, Korean, Thai, Chinese Simplified, and Chinese Traditional...

Reddit - Machine Learning · 1 min ·
Claude Mythos and Project Glasswing: why an AI superhacker has the tech world on alert
Llms

Claude Mythos and Project Glasswing: why an AI superhacker has the tech world on alert

A new AI model could automate the process of searching for cybersecurity bugs and flaws – for better or worse.

AI Tools & Products · 5 min ·
Gemini could take a 'proactive' approach with leaked 'Your Day' feature
Llms

Gemini could take a 'proactive' approach with leaked 'Your Day' feature

This feature could leverage your apps in a way that might feel familiar.

AI Tools & Products · 5 min ·
I ditched my paper planner for Gemini Live — and it solved the one professional problem I couldn't fix
Llms

I ditched my paper planner for Gemini Live — and it solved the one professional problem I couldn't fix

Can Gemini Live replace a physical planner? Tom's Guide AI Editor Amanda Caswell ditched her notebook for Google’s voice AI. Here’s how i...

AI Tools & Products · 8 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime