[2602.12708] Mixture of Predefined Experts: Maximizing Data Usage on Vertical Federated Learning

[2602.12708] Mixture of Predefined Experts: Maximizing Data Usage on Vertical Federated Learning

arXiv - Machine Learning 4 min read Article

Summary

The paper introduces Split-MoPE, a novel framework for Vertical Federated Learning that maximizes data usage by integrating predefined experts, enhancing performance in privacy-sensitive domains.

Why It Matters

As Vertical Federated Learning becomes increasingly vital in sectors like finance and healthcare, addressing the challenges of data misalignment is crucial. Split-MoPE offers a robust solution that enhances data utilization and model performance while maintaining privacy, making it relevant for researchers and practitioners in machine learning and data science.

Key Takeaways

  • Split-MoPE integrates Split Learning with a Mixture of Predefined Experts to optimize data usage.
  • The framework is designed to handle sample misalignment, a common issue in real-world federated learning scenarios.
  • Split-MoPE achieves superior performance with reduced communication needs compared to traditional methods.
  • It provides interpretability by quantifying each participant's contribution to predictions.
  • Extensive evaluations show its effectiveness on both vision and tabular datasets.

Computer Science > Machine Learning arXiv:2602.12708 (cs) [Submitted on 13 Feb 2026] Title:Mixture of Predefined Experts: Maximizing Data Usage on Vertical Federated Learning Authors:Jon Irureta, Gorka Azkune, Jon Imaz, Aizea Lojo, Javier Fernandez-Marques View a PDF of the paper titled Mixture of Predefined Experts: Maximizing Data Usage on Vertical Federated Learning, by Jon Irureta and 4 other authors View PDF HTML (experimental) Abstract:Vertical Federated Learning (VFL) has emerged as a critical paradigm for collaborative model training in privacy-sensitive domains such as finance and healthcare. However, most existing VFL frameworks rely on the idealized assumption of full sample alignment across participants, a premise that rarely holds in real-world scenarios. To bridge this gap, this work introduces Split-MoPE, a novel framework that integrates Split Learning with a specialized Mixture of Predefined Experts (MoPE) architecture. Unlike standard Mixture of Experts (MoE), where routing is learned dynamically, MoPE uses predefined experts to process specific data alignments, effectively maximizing data usage during both training and inference without requiring full sample overlap. By leveraging pretrained encoders for target data domains, Split-MoPE achieves state-of-the-art performance in a single communication round, significantly reducing the communication footprint compared to multi-round end-to-end training. Furthermore, unlike existing proposals that address sam...

Related Articles

A Machine Learning Engineer Thought He Was Safe From AI Layoffs. Then He Got Some Depressing News
Machine Learning

A Machine Learning Engineer Thought He Was Safe From AI Layoffs. Then He Got Some Depressing News

AI News - General · 4 min ·
UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
When AI training wheels help and hinder learning
Machine Learning

When AI training wheels help and hinder learning

AI News - General · 6 min ·
Sam Altman's Coworkers Say He Can Barely Code and Misunderstands Basic Machine Learning Concepts
Machine Learning

Sam Altman's Coworkers Say He Can Barely Code and Misunderstands Basic Machine Learning Concepts

AI News - General · 2 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime