[2407.17491] Robust Adaptation of Foundation Models with Black-Box Visual Prompting

[2407.17491] Robust Adaptation of Foundation Models with Black-Box Visual Prompting

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2407.17491: Robust Adaptation of Foundation Models with Black-Box Visual Prompting

Computer Science > Computer Vision and Pattern Recognition arXiv:2407.17491 (cs) [Submitted on 4 Jul 2024 (v1), last revised 1 Apr 2026 (this version, v3)] Title:Robust Adaptation of Foundation Models with Black-Box Visual Prompting Authors:Changdae Oh, Gyeongdeok Seo, Geunyoung Jung, Zhi-Qi Cheng, Hosik Choi, Jiyoung Jung, Kyungwoo Song View a PDF of the paper titled Robust Adaptation of Foundation Models with Black-Box Visual Prompting, by Changdae Oh and 6 other authors View PDF HTML (experimental) Abstract:With a surge of large-scale pre-trained models, parameter-efficient transfer learning (PETL) of large models has garnered significant attention. While promising, they commonly rely on two optimistic assumptions: 1) full access to the parameters of a PTM, and 2) sufficient memory capacity to cache all intermediate activations for gradient computation. However, in most real-world applications, PTMs serve as black-box APIs or proprietary software without full parameter accessibility. Besides, it is hard to meet a large memory requirement for modern PTMs. This work proposes black-box visual prompting (BlackVIP), which efficiently adapts the PTMs without knowledge of their architectures or parameters. BlackVIP has two components: 1) Coordinator and 2) simultaneous perturbation stochastic approximation with gradient correction (SPSA-GC). The Coordinator designs input-dependent visual prompts, which allow the target PTM to adapt in the wild. SPSA-GC efficiently estimates th...

Originally published on April 03, 2026. Curated by AI News.

Related Articles

Llms

Intel LLM-Scaler vllm-0.14.0-b8.2 released with official Arc Pro B70 support

submitted by /u/Fcking_Chuck [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
Llms

What was the biggest thing to happen in the field of AI?

I personally think it’s either AlphaGo or ChatGPT. AlphaGo showed to the whole world that AIs can be better than its creators in an area ...

Reddit - Artificial Intelligence · 1 min ·
Llms

Training-time intervention yields 63.4% blind-pair human preference at matched val-loss (1.2B params, 320 judgments, p = 1.98 × 10⁻⁵) [R]

TL;DR. I ran a blind A/B preference evaluation between two 1.2B-parameter LMs trained on identical data (same order, same seed, 30K steps...

Reddit - Machine Learning · 1 min ·
I tried Gemini, ChatGPT, and Claude for a month on Android, and I have a clear winner for you
Llms

I tried Gemini, ChatGPT, and Claude for a month on Android, and I have a clear winner for you

The ultimate Android AI showdown

AI Tools & Products · 5 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime