On the Necessity of Collaboration for Online Model Selection with Decentralized Data
About
We consider online model selection with decentralized data over $M$ clients, and study the necessity of collaboration among clients. Previous work proposed various federated algorithms without demonstrating their necessity,while we answer the question from a novel perspective of computational constraints. We prove lower bounds on the regret, and propose a federated algorithm and analyze the upper bound.Our results show (i) collaboration is unnecessary in the absence of computational constraints on clients; (ii) collaboration is necessary if the computational cost on each client is limited to $o(K)$, where $K$ is the number of candidate hypothesis spaces. We clarify the unnecessary nature of collaboration in previous federated algorithms for distributed online multi-kernel learning,and improve the regret bounds at a smaller computational and communication cost. Our algorithm relies on three new techniques including an improved Bernstein's inequality for martingale, a federated online mirror descent framework, and decoupling model selection and prediction, which might be of independent interest.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Regression | California Housing (CH) (test) | MSE0.0237 | 52 | |
| Regression | elevators (test) | RMSE0.0032 | 19 | |
| Regression | Slice (test) | MSE0.057 | 7 | |
| Regression | bank (test) | MSE0.0192 | 3 | |
| Regression | TomsHardware (test) | MSE5.90e-4 | 3 | |
| Regression | Twitter (test) | MSE1.00e-4 | 3 | |
| Regression | ailerons (test) | MSE0.0043 | 3 | |
| Online Model Selection with Decentralized Data | OMS-DecD (Theoretical) | -- | 2 |