Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Learning Models with Uniform Performance via Distributionally Robust Optimization

About

A common goal in statistics and machine learning is to learn models that can perform well against distributional shifts, such as latent heterogeneous subpopulations, unknown covariate shifts, or unmodeled temporal effects. We develop and analyze a distributionally robust stochastic optimization (DRO) framework that learns a model providing good performance against perturbations to the data-generating distribution. We give a convex formulation for the problem, providing several convergence guarantees. We prove finite-sample minimax upper and lower bounds, showing that distributional robustness sometimes comes at a cost in convergence rates. We give limit theorems for the learned parameters, where we fully specify the limiting distribution so that confidence intervals can be computed. On real tasks including generalizing to unknown subpopulations, fine-grained recognition, and providing good tail performance, the distributionally robust approach often exhibits improved performance.

John Duchi, Hongseok Namkoong• 2018

Related benchmarks

TaskDatasetResultRank
Image ClassificationWaterbirds (test)
Worst-Group Accuracy75.5
112
ClassificationCelebA (test)
Average Accuracy95.1
92
Natural Language InferenceMultiNLI (test)--
81
ClassificationCheXpert (test)--
48
ClassificationCivilComments (test)
Worst-case Accuracy68.7
47
Group RobustnessCivilComments-WILDS (test)
WG Accuracy57.1
40
Group RobustnessCheXpert (test)
WGA37.1
22
Text ClassificationMultiNLI (test)
WGA63
18
Toxicity DetectionCivilComments (test)
WGA62.9
14
Medical Image ClassificationCheXpert (test)
WGA0.502
8
Showing 10 of 10 rows

Other info

Follow for update