Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Prompting for Multimodal Hateful Meme Classification

About

Hateful meme classification is a challenging multimodal task that requires complex reasoning and contextual background knowledge. Ideally, we could leverage an explicit external knowledge base to supplement contextual and cultural information in hateful memes. However, there is no known explicit external knowledge base that could provide such hate speech contextual information. To address this gap, we propose PromptHate, a simple yet effective prompt-based model that prompts pre-trained language models (PLMs) for hateful meme classification. Specifically, we construct simple prompts and provide a few in-context examples to exploit the implicit knowledge in the pre-trained RoBERTa language model for hateful meme classification. We conduct extensive experiments on two publicly available hateful and offensive meme datasets. Our experimental results show that PromptHate is able to achieve a high AUC of 90.96, outperforming state-of-the-art baselines on the hateful meme classification task. We also perform fine-grained analyses and case studies on various prompt settings and demonstrate the effectiveness of the prompts on hateful meme classification.

Rui Cao, Roy Ka-Wei Lee, Wen-Haw Chong, Jing Jiang• 2023

Related benchmarks

TaskDatasetResultRank
Hateful Meme DetectionHateful Memes (test)
AUROC0.8145
67
Hateful meme classificationHarM (test)
AUC90.96
31
Harmful Meme DetectionFHM
Accuracy72.2
29
Harmful Meme DetectionMAMI
Accuracy70.4
19
Harmful Meme DetectionToxiCN
Accuracy76.04
16
ClassificationFHM
Accuracy72.98
16
ClassificationMAMI
Accuracy70.31
16
Hateful meme classificationHarMeme (test)
Accuracy84.5
15
Hateful Meme DetectionHarM
AUC87.51
12
Hateful Meme DetectionFHM
AUC76.76
12
Showing 10 of 23 rows

Other info

Code

Follow for update