Beyond Static Benchmarks: Synthesizing Harmful Content via Persona-based Simulation for Robust Evaluation

arXiv cs.CL / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that static harmful-content detection benchmarks are limited by low scalability and diversity, and they may be distorted by contamination from web-scale pretraining data.
  • It proposes a framework that uses persona-guided LLM agents to synthesize harmful content, combining demographic identities and topical interests with situational harmful strategies to simulate realistic harmful interactions.
  • The framework is evaluated along three axes—harmfulness, challenge level, and diversity—using both human assessments and LLM-based evaluations.
  • Results indicate a high harmful generation success rate and show that the synthetic scenarios are significantly more difficult for multiple existing detection systems to identify than scenarios from current benchmarks.
  • The authors report that the generated content achieves linguistic and topical diversity comparable to human-curated datasets, positioning the approach as a robust stress-testing tool for detection systems.

Abstract

Static benchmarks for harmful content detection face limitations in scalability and diversity, and may also be affected by contamination from web-scale pre-training corpora. To address these issues, we propose a framework for synthesizing harmful content, leveraging persona-guided large language model (LLM) agents. Our approach constructs two-dimensional user personas by integrating demographic identities and topical interests with situational harmful strategies, enabling the simulation of diverse and contextually grounded harmful interactions. We evaluate the framework along three dimensions: harmfulness, challenge level, and diversity. Both human and LLM-based evaluations confirm that our framework achieves a high harmful generation success rate. Experiments across multiple detection systems reveal that our synthetic scenarios are more challenging to detect than those in existing benchmarks. Furthermore, a multi-faceted analysis confirms that our approach achieves linguistic and topical diversity comparable to human-curated datasets, establishing our framework as an effective tool for robust stress-testing of harmful content detection systems.