Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Mixture of Predefined Experts: Maximizing Data Usage on Vertical Federated Learning

About

Vertical Federated Learning (VFL) has emerged as a critical paradigm for collaborative model training in privacy-sensitive domains such as finance and healthcare. However, most existing VFL frameworks rely on the idealized assumption of full sample alignment across participants, a premise that rarely holds in real-world scenarios. To bridge this gap, this work introduces Split-MoPE, a novel framework that integrates Split Learning with a specialized Mixture of Predefined Experts (MoPE) architecture. Unlike standard Mixture of Experts (MoE), where routing is learned dynamically, MoPE uses predefined experts to process specific data alignments, effectively maximizing data usage during both training and inference without requiring full sample overlap. By leveraging pretrained encoders for target data domains, Split-MoPE achieves state-of-the-art performance in a single communication round, significantly reducing the communication footprint compared to multi-round end-to-end training. Furthermore, unlike existing proposals that address sample misalignment, this novel architecture provides inherent robustness against malicious or noisy participants and offers per-sample interpretability by quantifying each collaborator's contribution to each prediction. Extensive evaluations on vision (CIFAR-10/100) and tabular (Breast Cancer Wisconsin) datasets demonstrate that Split-MoPE consistently outperforms state-of-the-art systems such as LASER and Vertical SplitNN, particularly in challenging scenarios with high data missingness.

Jon Irureta, Gorka Azkune, Jon Imaz, Aizea Lojo, Javier Fernandez-Marques• 2026

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy81.78
3518
Image ClassificationCIFAR-10 (test)
Accuracy95.63
3381
Image ClassificationCIFAR10 0.1--
11
Image ClassificationCIFAR-10 p_miss = 0
Error (Noise Level 1)0.35
3
Image ClassificationCIFAR-100 p_miss = 0
Error Rate (Noise Level 1)3.15
3
Image ClassificationCIFAR-100 p_miss = 0.1
Miss Rate (L1)3.36
3
Showing 6 of 6 rows

Other info

Follow for update