Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Identity Mappings in Deep Residual Networks

About

Deep residual networks have emerged as a family of extremely deep architectures showing compelling accuracy and nice convergence behaviors. In this paper, we analyze the propagation formulations behind the residual building blocks, which suggest that the forward and backward signals can be directly propagated from one block to any other block, when using identity mappings as the skip connections and after-addition activation. A series of ablation experiments support the importance of these identity mappings. This motivates us to propose a new residual unit, which makes training easier and improves generalization. We report improved results using a 1001-layer ResNet on CIFAR-10 (4.62% error) and CIFAR-100, and a 200-layer ResNet on ImageNet. Code is available at: https://github.com/KaimingHe/resnet-1k-layers

Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun• 2016

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy75.67
3518
Image ClassificationCIFAR-10 (test)
Accuracy94.54
3381
Object DetectionCOCO 2017 (val)
AP38.5
2454
Image ClassificationImageNet-1k (val)--
1453
Image ClassificationImageNet (val)
Top-1 Acc23.77
1206
Semantic segmentationCityscapes (test)--
1145
Image ClassificationCIFAR-10 (test)--
906
Image ClassificationMNIST (test)--
882
Image ClassificationCIFAR-100 (val)--
661
Image ClassificationCIFAR-100--
622
Showing 10 of 43 rows

Other info

Code

Follow for update