Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Do Deep Nets Really Need to be Deep?

About

Currently, deep neural networks are the state of the art on problems such as speech recognition and computer vision. In this extended abstract, we show that shallow feed-forward networks can learn the complex functions previously learned by deep nets and achieve accuracies previously only achievable with deep models. Moreover, in some cases the shallow neural nets can learn these deep functions using a total number of parameters similar to the original deep model. We evaluate our method on the TIMIT phoneme recognition task and are able to train shallow fully-connected nets that perform similarly to complex, well-engineered, deep convolutional architectures. Our success in training shallow neural nets to mimic deeper models suggests that there probably exist better algorithms for training shallow feed-forward nets than those currently available.

Lei Jimmy Ba, Rich Caruana• 2013

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy74.08
3518
Image ClassificationImageNet-1k (val)--
1453
Image ClassificationMNIST (test)
Accuracy99.49
882
Image ClassificationImageNet-1K
Top-1 Acc56.58
836
Image ClassificationCIFAR-100
Top-1 Accuracy73.08
622
Image ClassificationCIFAR-10--
507
Image ClassificationMNIST--
395
Image ClassificationTinyImageNet (test)--
366
Image ClassificationTiny-ImageNet
Accuracy34.2
227
Image ClassificationTiny-ImageNet
Top-1 Accuracy60.07
143
Showing 10 of 18 rows

Other info

Follow for update