DSIPA: Detecting LLM-Generated Texts via Sentiment-Invariant Patterns Divergence Analysis

arXiv cs.AI / 4/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DSIPA, a training-free, zero-shot framework to detect LLM-generated text by analyzing how sentiment patterns change under controlled stylistic variations.
  • DSIPA is designed to be robust to adversarial perturbations, paraphrasing attacks, and domain shifts, avoiding common requirements such as access to model parameters or large labeled datasets.
  • It operates in a black-box setting using two unsupervised metrics—sentiment distribution consistency and sentiment distribution preservation—to capture differences between typically emotion-stable LLM outputs and more affectively diverse human writing.
  • Experiments across multiple state-of-the-art proprietary and open-source models (e.g., GPT-5.2, Gemini-1.5-pro, Claude-3, LLaMa-3.3) and five domains show improvements in F1 scores by up to 49.89% versus baseline detection methods.
  • The authors report strong cross-domain generalization and resilience to adversarial conditions, presenting an interpretable behavioral signal for secure content identification as LLM capabilities evolve.

Abstract

The rapid advancement of large language models (LLMs) presents new security challenges, particularly in detecting machine-generated text used for misinformation, impersonation, and content forgery. Most existing detection approaches struggle with robustness against adversarial perturbation, paraphrasing attacks, and domain shifts, often requiring restrictive access to model parameters or large labeled datasets. To address this, we propose DSIPA, a novel training-free framework that detects LLM-generated content by quantifying sentiment distributional stability under controlled stylistic variation. It is based on the observation that LLMs typically exhibit more emotionally consistent outputs, while human-written texts display greater affective variation. Our framework operates in a zero-shot, black-box manner, leveraging two unsupervised metrics, sentiment distribution consistency and sentiment distribution preservation, to capture these intrinsic behavioral asymmetries without the need for parameter updates or probability access. Extensive experiments are conducted on state-of-the-art proprietary and open-source models, including GPT-5.2, Gemini-1.5-pro, Claude-3, and LLaMa-3.3. Evaluations on five domains, such as news articles, programming code, student essays, academic papers, and community comments, demonstrate that DSIPA improves F1 detection scores by up to 49.89% over baseline methods. The framework exhibits superior generalizability across domains and strong resilience to adversarial conditions, providing a robust and interpretable behavioral signal for secure content identification in the evolving LLM landscape.