Inconsistency, Instability, and Generalization Gap of Deep Neural Network Training
About
As deep neural networks are highly expressive, it is important to find solutions with small generalization gap (the difference between the performance on the training data and unseen data). Focusing on the stochastic nature of training, we first present a theoretical analysis in which the bound of generalization gap depends on what we call inconsistency and instability of model outputs, which can be estimated on unlabeled data. Our empirical study based on this analysis shows that instability and inconsistency are strongly predictive of generalization gap in various settings. In particular, our finding indicates that inconsistency is a more reliable indicator of generalization gap than the sharpness of the loss landscape. Furthermore, we show that algorithmic reduction of inconsistency leads to superior performance. The results also provide a theoretical basis for existing methods such as co-distillation and ensemble.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | Food-101 | -- | 494 | |
| Image Classification | CIFAR-10 standard (train=4K, unlabeled=41K) | Error Rate3.95 | 6 | |
| Generalization Gap Prediction | CIFAR-10 | -- | 6 | |
| Generalization Gap Prediction | ImageNet | Gap Prediction Error13 | 5 | |
| Generalization Gap Prediction | Food101 | Generalization Gap Prediction Error10 | 5 | |
| Generalization Gap Prediction | Stanford Dogs | Generalization Gap Prediction Error27 | 5 | |
| Generalization Gap Prediction | MNLI | Gap Prediction Error0.09 | 5 | |
| Generalization Gap Prediction | MNLI Case 8 | Gap Prediction Error15 | 5 | |
| Generalization Gap Prediction | QNLI Case 9 | Gap Prediction Error0.18 | 5 | |
| Generalization Gap Prediction | Food101 Distillation | Generalization Gap Prediction Error0.19 | 5 |