Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

CoLES: Contrastive Learning for Event Sequences with Self-Supervision

About

We address the problem of self-supervised learning on discrete event sequences generated by real-world users. Self-supervised learning incorporates complex information from the raw data in low-dimensional fixed-length vector representations that could be easily applied in various downstream machine learning tasks. In this paper, we propose a new method "CoLES", which adapts contrastive learning, previously used for audio and computer vision domains, to the discrete event sequences domain in a self-supervised setting. We deployed CoLES embeddings based on sequences of transactions at the large European financial services company. Usage of CoLES embeddings significantly improves the performance of the pre-existing models on downstream tasks and produces significant financial gains, measured in hundreds of millions of dollars yearly. We also evaluated CoLES on several public event sequences datasets and showed that CoLES representations consistently outperform other methods on different downstream tasks.

Dmitrii Babaev, Ivan Kireev, Nikita Ovsov, Mariya Ivanova, Gleb Gusev, Ivan Nazarov, Alexander Tuzhilin• 2020

Related benchmarks

TaskDatasetResultRank
ClassificationInternal Bank Data (test)
AUC0.905
12
Binary ClassificationGender 2019 (test)
AUC0.877
12
Multi-class classificationAge sberbank-sirius-lesson (test)
Accuracy63.7
12
Age PredictionAge
Accuracy64.5
12
ClassificationRosbank
AUC0.835
12
Gender PredictionGender
AUC88.8
10
ClassificationDataFusion
AUC0.738
10
Age ClassificationPrivate Dataset
Accuracy74.3
6
Gender ClassificationPrivate Dataset
AUC89.8
6
RegressionPrivate Dataset
MAE1.14e+4
6
Showing 10 of 12 rows

Other info

Follow for update