Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

UKD: Debiasing Conversion Rate Estimation via Uncertainty-regularized Knowledge Distillation

About

In online advertising, conventional post-click conversion rate (CVR) estimation models are trained using clicked samples. However, during online serving the models need to estimate for all impression ads, leading to the sample selection bias (SSB) issue. Intuitively, providing reliable supervision signals for unclicked ads is a feasible way to alleviate the SSB issue. This paper proposes an uncertainty-regularized knowledge distillation (UKD) framework to debias CVR estimation via distilling knowledge from unclicked ads. A teacher model learns click-adaptive representations and produces pseudo-conversion labels on unclicked ads as supervision signals. Then a student model is trained on both clicked and unclicked ads with knowledge distillation, performing uncertainty modeling to alleviate the inherent noise in pseudo-labels. Experiments on billion-scale datasets show that UKD outperforms previous debiasing methods. Online results verify that UKD achieves significant improvements.

Zixuan Xu, Penghui Wei, Weimin Zhang, Shaoguo Liu, Liang Wang, Bo Zheng• 2022

Related benchmarks

TaskDatasetResultRank
RecommendationIndustrial Dataset Taobao (test)
HR@30.5061
9
RecommendationTaobao-MM (test)
HR@346.1
9
Showing 2 of 2 rows

Other info

Follow for update