Relative Density-Ratio Estimation for Robust Distribution Comparison
About
Divergence estimators based on direct approximation of density-ratios without going through separate approximation of numerator and denominator densities have been successfully applied to machine learning tasks that involve distribution comparison such as outlier detection, transfer learning, and two-sample homogeneity test. However, since density-ratio functions often possess high fluctuation, divergence estimation is still a challenging task in practice. In this paper, we propose to use relative divergences for distribution comparison, which involves approximation of relative density-ratios. Since relative density-ratios are always smoother than corresponding ordinary density-ratios, our proposed method is favorable in terms of the non-parametric convergence speed. Furthermore, we show that the proposed divergence estimator has asymptotic variance independent of the model complexity under a parametric setup, implying that the proposed estimator hardly overfits even with complex models. Through experiments, we demonstrate the usefulness of the proposed approach.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | MNIST-10 (test) | Test Accuracy74.94 | 19 | |
| Image Classification | MNIST CS 100 (test) | Mean Accuracy76.53 | 19 | |
| Image Classification | Shift-MNIST (test) | Accuracy80.15 | 15 | |
| Image Classification | CIFAR-20 case (iii) (test) | Accuracy58.02 | 12 | |
| Image Classification | MNIST Label Noise 0.4 (test) | Mean Accuracy81.19 | 12 | |
| Image Classification | Color-MNIST case (iii) (test) | Accuracy39.28 | 12 | |
| Image Classification | MNIST Label Noise 0.2 (test) | Mean Accuracy82.7 | 12 | |
| Image Classification | MNIST case (iii) (test) | Accuracy81.73 | 12 | |
| Image Classification | MNIST case (iv) (test) | Accuracy77.11 | 12 | |
| Image Classification | CIFAR-20 Label Noise 0.2 (test) | Mean Accuracy50.13 | 7 |