[2510.06162] TabPFN-Wide: Continued Pre-Training for Extreme Feature Counts

[2510.06162] TabPFN-Wide: Continued Pre-Training for Extreme Feature Counts

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2510.06162: TabPFN-Wide: Continued Pre-Training for Extreme Feature Counts

Computer Science > Machine Learning arXiv:2510.06162 (cs) [Submitted on 7 Oct 2025 (v1), last revised 29 Mar 2026 (this version, v2)] Title:TabPFN-Wide: Continued Pre-Training for Extreme Feature Counts Authors:Christopher Kolberg, Jules Kreuer, Jonas Huurdeman, Sofiane Ouaari, Katharina Eggensperger, Nico Pfeifer View a PDF of the paper titled TabPFN-Wide: Continued Pre-Training for Extreme Feature Counts, by Christopher Kolberg and 4 other authors View PDF HTML (experimental) Abstract:Revealing novel insights from the relationship between molecular measurements and pathology remains a very impactful application of machine learning in biomedicine. Data in this domain typically contain only a few observations but thousands of potentially noisy features, posing challenges for conventional tabular machine learning approaches. While prior-data fitted networks emerge as foundation models for predictive tabular data tasks, they are currently not suited to handle large feature counts (>500). Although feature reduction enables their application, it hinders feature importance analysis. We propose a strategy that extends existing models through continued pre-training on synthetic data sampled from a customized prior. The resulting model, TabPFN-Wide, matches or exceeds its base model's performance, while exhibiting improved robustness to noise. It seamlessly scales beyond 30,000 categorical and continuous features, regardless of noise levels, while maintaining inherent interpretabi...

Originally published on March 31, 2026. Curated by AI News.

Related Articles

Llms

OTHER AI PLATFORMS BETTER THAN CHATGPT (OPEN AI)

Share your thoughts submitted by /u/InnerNeedleworker347 [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
Llms

Seeking Critique on Research Approach to Open Set Recognition (Novelty Detection) [R]

Hey guys, I'm an independent researcher working on a project that tries to address a very specific failure mode in LLMs and embedding bas...

Reddit - Machine Learning · 1 min ·
Google rolls out a native Gemini app for Mac | TechCrunch
Llms

Google rolls out a native Gemini app for Mac | TechCrunch

You can share anything on their screen with Gemini to get help with what they're looking at in the moment, including local files.

TechCrunch - AI · 3 min ·
Llms

Coherence under Constraint

I’ve been running some small experiments forcing LLMs into contradictions they can’t resolve. What surprised me wasn’t that they fail—it’...

Reddit - Artificial Intelligence · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime