Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Improving Collaborative Metric Learning with Efficient Negative Sampling

About

Distance metric learning based on triplet loss has been applied with success in a wide range of applications such as face recognition, image retrieval, speaker change detection and recently recommendation with the CML model. However, as we show in this article, CML requires large batches to work reasonably well because of a too simplistic uniform negative sampling strategy for selecting triplets. Due to memory limitations, this makes it difficult to scale in high-dimensional scenarios. To alleviate this problem, we propose here a 2-stage negative sampling strategy which finds triplets that are highly informative for learning. Our strategy allows CML to work effectively in terms of accuracy and popularity bias, even when the batch size is an order of magnitude smaller than what would be needed with the default uniform sampling. We demonstrate the suitability of the proposed strategy for recommendation and exhibit consistent positive results across various datasets.

Viet-Anh Tran, Romain Hennequin, Jimena Royo-Letelier, Manuel Moussallam• 2019

Related benchmarks

TaskDatasetResultRank
RecommendationMovieLens 1M (test)
Recall@34.84
34
RecommendationMovieLens 10M
Precision@30.1647
16
RecommendationSteam-200k 1.0 (test)
P@325.2
16
RecommendationCiteULike
Precision@36.4
16
Showing 4 of 4 rows

Other info

Follow for update