Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Improving Generalizability in Implicitly Abusive Language Detection with Concept Activation Vectors

About

Robustness of machine learning models on ever-changing real-world data is critical, especially for applications affecting human well-being such as content moderation. New kinds of abusive language continually emerge in online discussions in response to current events (e.g., COVID-19), and the deployed abuse detection systems should be updated regularly to remain accurate. In this paper, we show that general abusive language classifiers tend to be fairly reliable in detecting out-of-domain explicitly abusive utterances but fail to detect new types of more subtle, implicit abuse. Next, we propose an interpretability technique, based on the Testing Concept Activation Vector (TCAV) method from computer vision, to quantify the sensitivity of a trained model to the human-defined concepts of explicit and implicit abusive language, and use that to explain the generalizability of the model on new data, in this case, COVID-related anti-Asian hate speech. Extending this technique, we introduce a novel metric, Degree of Explicitness, for a single instance and show that the new metric is beneficial in suggesting out-of-domain unlabeled examples to effectively enrich the training data with informative, implicitly abusive texts.

Isar Nejadgholi, Kathleen C. Fraser, Svetlana Kiritchenko• 2022

Related benchmarks

TaskDatasetResultRank
Abusive language detectionEA (test)
F1 Score61
4
Abusive language detectionCH (test)
F1 Score73
4
Abusive language detectionWiki (test)
F1 Score82
4
Showing 3 of 3 rows

Other info

Code

Follow for update