Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Deep Representations with Probabilistic Knowledge Transfer

About

Knowledge Transfer (KT) techniques tackle the problem of transferring the knowledge from a large and complex neural network into a smaller and faster one. However, existing KT methods are tailored towards classification tasks and they cannot be used efficiently for other representation learning tasks. In this paper a novel knowledge transfer technique, that is capable of training a student model that maintains the same amount of mutual information between the learned representation and a set of (possible unknown) labels as the teacher model, is proposed. Apart from outperforming existing KT techniques, the proposed method allows for overcoming several limitations of existing methods providing new insight into KT as well as novel KT applications, ranging from knowledge transfer from handcrafted feature extractors to {cross-modal} KT from the textual modality into the representation extracted from the visual modality of the data.

Nikolaos Passalis, Anastasios Tefas• 2018

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy74.69
3518
Image ClassificationCIFAR100 (test)
Top-1 Accuracy76.01
377
Image ClassificationCIFAR100 (test)
Test Accuracy75.51
147
Image ClassificationCIFAR100
Average Accuracy74.54
121
Image ClassificationCIFAR-100
Nominal Accuracy74.69
116
Image GenerationCIFAR-100 (10% data)--
41
Video ClassificationKinetics-400 v1 (val)
Top-1 Acc68.37
35
Image GenerationCIFAR-10 (10% data)--
35
Action ClassificationActivityNet (val)
Top-1 Acc35.4
30
Video RetrievalUCF51
Recall@10.612
27
Showing 10 of 16 rows

Other info

Follow for update