[2603.00478] Benchmarking Few-shot Transferability of Pre-trained Models with Improved Evaluation Protocols

[2603.00478] Benchmarking Few-shot Transferability of Pre-trained Models with Improved Evaluation Protocols

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2603.00478: Benchmarking Few-shot Transferability of Pre-trained Models with Improved Evaluation Protocols

Computer Science > Machine Learning arXiv:2603.00478 (cs) [Submitted on 28 Feb 2026] Title:Benchmarking Few-shot Transferability of Pre-trained Models with Improved Evaluation Protocols Authors:Xu Luo, Ji Zhang, Lianli Gao, Heng Tao Shen, Jingkuan Song View a PDF of the paper titled Benchmarking Few-shot Transferability of Pre-trained Models with Improved Evaluation Protocols, by Xu Luo and 3 other authors View PDF HTML (experimental) Abstract:Few-shot transfer has been revolutionized by stronger pre-trained models and improved adaptation this http URL, there lacks a unified, rigorous evaluation protocol that is both challenging and realistic for real-world usage. In this work, we establish FEWTRANS, a comprehensive benchmark containing 10 diverse datasets, and propose the Hyperparameter Ensemble (HPE) protocol to overcome the "validation set illusion" in data-scarce regimes. Our empirical findings demonstrate that the choice of pre-trained model is the dominant factor for performance, while many sophisticated transfer methods offer negligible practical advantages over a simple full-parameter fine-tuning baseline. To explain this surprising effectiveness, we provide an in-depth mechanistic analysis showing that full fine-tuning succeeds via distributed micro-adjustments and more flexible reshaping of high-level semantic presentations without suffering from overfitting. Additionally, we quantify the performance collapse of multimodal models in specialized domains as a resul...

Originally published on March 03, 2026. Curated by AI News.

Related Articles

AI chip startup Rebellions raises $400 million at $2.3B valuation in pre-IPO round | TechCrunch
Machine Learning

AI chip startup Rebellions raises $400 million at $2.3B valuation in pre-IPO round | TechCrunch

The startup, which is planning to go public later this year, designs chips specifically for AI inference, another challenger to Nvidia's ...

TechCrunch - AI · 4 min ·
Llms

CLI for Google AI Search (gai.google) — run AI-powered code/tech searches headlessly from your terminal

Google AI (gai.google) gives Gemini-powered answers for technical queries — think AI-enhanced search with code understanding. I built a C...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

Big increase in the amount of people using AI to write their replies with AI

I find it interesting that we’ve all randomly decided to use the “-“ more often recently on reddit, and everyone’s grammar has drasticall...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

[D] MXFP8 GEMM: Up to 99% of cuBLAS performance using CUDA + PTX

New blog post by Daniel Vega-Myhre (Meta/PyTorch) illustrating GEMM design for FP8, including deep-dives into all the constraints and des...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime