Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

StableDR: Stabilized Doubly Robust Learning for Recommendation on Data Missing Not at Random

About

In recommender systems, users always choose the favorite items to rate, which leads to data missing not at random and poses a great challenge for unbiased evaluation and learning of prediction models. Currently, the doubly robust (DR) methods have been widely studied and demonstrate superior performance. However, in this paper, we show that DR methods are unstable and have unbounded bias, variance, and generalization bounds to extremely small propensities. Moreover, the fact that DR relies more on extrapolation will lead to suboptimal performance. To address the above limitations while retaining double robustness, we propose a stabilized doubly robust (StableDR) learning approach with a weaker reliance on extrapolation. Theoretical analysis shows that StableDR has bounded bias, variance, and generalization error bound simultaneously under inaccurate imputed errors and arbitrarily small propensities. In addition, we propose a novel learning approach for StableDR that updates the imputation, propensity, and prediction models cyclically, achieving more stable and accurate predictions. Extensive experiments show that our approaches significantly outperform the existing methods.

Haoxuan Li, Chunyuan Zheng, Peng Wu• 2022

Related benchmarks

TaskDatasetResultRank
Reward ModelingHelpSteer (test)
MAE0.318
48
Safety EvaluationHarmBench--
42
Safety EvaluationWildGuardMix
Safety Score0.8534
22
Reward ModelingUltraFeedback (test)
MAE0.272
21
Reward ModelingPKU-SafeRLHF (test)
MAE0.1771
19
Safety EvaluationDAN
Safety Score (DAN)0.782
18
Safety AlignmentStrongREJECT--
18
Rating PredictionMusic unbiased (test)
AUC68.7
12
Rating PredictionCoat unbiased (test)
AUC0.719
12
Rating PredictionKuaiRec unbiased (test)
AUC76.4
12
Showing 10 of 11 rows

Other info

Follow for update