Explaining Preferences with Shapley Values
About
While preference modelling is becoming one of the pillars of machine learning, the problem of preference explanation remains challenging and underexplored. In this paper, we propose \textsc{Pref-SHAP}, a Shapley value-based model explanation framework for pairwise comparison data. We derive the appropriate value functions for preference models and further extend the framework to model and explain \emph{context specific} information, such as the surface type in a tennis game. To demonstrate the utility of \textsc{Pref-SHAP}, we apply our method to a variety of synthetic and real-world datasets and show that richer and more insightful explanations can be obtained over the baseline.
Robert Hu, Siu Lun Chau, Jaime Ferrando Huertas, Dino Sejdinovic• 2022
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Preference Learning | Synthetic (test) | Test AUC98 | 4 | |
| Preference Learning | Chameleon (test) | Test AUC92 | 4 | |
| Preference Learning | Pokémon (test) | Test AUC86 | 4 | |
| Preference Learning | Tennis (test) | Test AUC0.58 | 4 | |
| Preference Learning | Website (test) | Test AUC0.66 | 2 |
Showing 5 of 5 rows