Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MentalBERT: Publicly Available Pretrained Language Models for Mental Healthcare

About

Mental health is a critical issue in modern society, and mental disorders could sometimes turn to suicidal ideation without adequate treatment. Early detection of mental disorders and suicidal ideation from social content provides a potential way for effective social intervention. Recent advances in pretrained contextualized language representations have promoted the development of several domain-specific pretrained models and facilitated several downstream applications. However, there are no existing pretrained language models for mental healthcare. This paper trains and release two pretrained masked language models, i.e., MentalBERT and MentalRoBERTa, to benefit machine learning for the mental healthcare research community. Besides, we evaluate our trained domain-specific models and several variants of pretrained language models on several mental disorder detection benchmarks and demonstrate that language representations pretrained in the target domain improve the performance of mental health detection tasks.

Shaoxiong Ji, Tianlin Zhang, Luna Ansari, Jie Fu, Prayag Tiwari, Erik Cambria• 2021

Related benchmarks

TaskDatasetResultRank
Depressive symptom classificationRESTOREx (test)
Macro F1 Score64.62
19
Stress DetectionSMM4H Task 8 Twitter Stress Detection 2022 (test)
Recall85
12
Binary ClassificationTherapist Q&A (test)
Kappa (κF1)0.515
12
Multilabel ClassificationTherapist Q&A (test)
Kappa (κF1)0.169
12
Multiclass ClassificationTherapist Q&A (test)
Kappa (κF1)0.257
8
Therapeutic Strategy ClassificationPST Strategy Classification Dataset (test)
Weighted F1 (PS)78
4
Binary ClassificationGpt4 0.5
Weighted F183.8
3
Binary ClassificationGpt4 0.7
Weighted F185.4
3
Binary ClassificationGpt4o 0.5
Weighted F183.2
3
Binary ClassificationGpt4o 0.7
Weighted F182.9
3
Showing 10 of 18 rows

Other info

Follow for update