Tool-MCoT: Tool Augmented Multimodal Chain-of-Thought for Content Safety Moderation

arXiv cs.CL / 4/9/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Tool-MCoT is presented as a tool-augmented multimodal chain-of-thought approach for content safety moderation, aimed at handling complex inputs across different media types.
  • The method fine-tunes a small language model (SLM) using tool-augmented chain-of-thought training data generated by larger LLMs to improve reasoning and moderation decisions.
  • Experiments reported in the paper show significant performance gains from the fine-tuned SLM compared with baselines, while maintaining practical moderation effectiveness.
  • A key efficiency contribution is that the model learns to call external tools selectively, improving the trade-off between moderation accuracy and inference latency/cost.
  • The work targets the scalability challenge of deploying LLM-based moderation systems by reducing computational overhead through SLM deployment with tool augmentation.

Abstract

The growth of online platforms and user content requires strong content moderation systems that can handle complex inputs from various media types. While large language models (LLMs) are effective, their high computational cost and latency present significant challenges for scalable deployment. To address this, we introduce Tool-MCoT, a small language model (SLM) fine-tuned for content safety moderation leveraging external framework. By training our model on tool-augmented chain-of-thought data generated by LLM, we demonstrate that the SLM can learn to effectively utilize these tools to improve its reasoning and decision-making. Our experiments show that the fine-tuned SLM achieves significant performance gains. Furthermore, we show that the model can learn to use these tools selectively, achieving a balance between moderation accuracy and inference efficiency by calling tools only when necessary.