Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Distributed learning of deep neural network over multiple agents

About

In domains such as health care and finance, shortage of labeled data and computational resources is a critical issue while developing machine learning algorithms. To address the issue of labeled data scarcity in training and deployment of neural network-based systems, we propose a new technique to train deep neural networks over several data sources. Our method allows for deep neural networks to be trained using data from multiple entities in a distributed fashion. We evaluate our algorithm on existing datasets and show that it obtains performance which is similar to a regular neural network trained on a single machine. We further extend it to incorporate semi-supervised learning when training with few labeled samples, and analyze any security concerns that may arise. Our algorithm paves the way for distributed training of deep neural networks in data sensitive applications when raw data may not be shared directly.

Otkrist Gupta, Ramesh Raskar• 2018

Related benchmarks

TaskDatasetResultRank
Membership Inference Attack DefenseCIFAR100 (test)
Loss (Series)0.69
60
Membership Inference Attack DefenseCIFAR10
AUC (Loss-Series)62
26
Membership InferenceTinyImageNet
Loss0.64
23
Membership Inference DefenseTinyImageNet (test)
AUC (Loss-Series)0.7
15
Defense against Membership Inference AttacksCIFAR10
Loss Series Score0.61
15
Membership Inference Attack DefenseCIFAR100 Half Case
Loss-Series AUC0.71
8
Membership Inference Attack DefenseCIFAR100 Pair Case
Loss-Series AUC0.7
8
Membership InferenceCIFAR10 Pair (test)
Loss10.23
8
Membership InferenceCIFAR10 Half (test)
Loss Series8.46
7
Membership Inference Attack DefenseTinyImageNet (Half)
Loss0.65
7
Showing 10 of 10 rows

Other info

Follow for update