Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Tool-MCoT: Tool Augmented Multimodal Chain-of-Thought for Content Safety Moderation

About

The growth of online platforms and user content requires strong content moderation systems that can handle complex inputs from various media types. While large language models (LLMs) are effective, their high computational cost and latency present significant challenges for scalable deployment. To address this, we introduce Tool-MCoT, a small language model (SLM) fine-tuned for content safety moderation leveraging external framework. By training our model on tool-augmented chain-of-thought data generated by LLM, we demonstrate that the SLM can learn to effectively utilize these tools to improve its reasoning and decision-making. Our experiments show that the fine-tuned SLM achieves significant performance gains. Furthermore, we show that the model can learn to use these tools selectively, achieving a balance between moderation accuracy and inference efficiency by calling tools only when necessary.

Shutong Zhang, Dylan Zhou, Yinxiao Liu, Yang Yang, Huiwen Luo, Wenfei Zou• 2026

Related benchmarks

TaskDatasetResultRank
Multimodal Content ModerationHatefulMemes
Accuracy80.8
4
Multimodal Content ModerationMMHS150K
Accuracy66.2
4
Multimodal Content ModerationUnsafeBench
Accuracy76.7
4
Showing 3 of 3 rows

Other info

Follow for update