Content Fuzzing for Escaping Information Cocoons on Digital Social Media

arXiv cs.CL / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that social-media “information cocoons” are reinforced when stance detection signals are used in recommendation/ranking to route users toward like-minded content.
  • It proposes a creator-focused approach to revise posts so they can reach beyond existing affinity clusters and expose users to more diverse viewpoints.
  • The authors introduce ContentFuzz, a confidence-guided fuzzing framework that uses an LLM to produce meaning-preserving rewrites while exploiting feedback from stance-detection models.
  • Across four stance-detection models, three datasets, and two languages, ContentFuzz is reported to successfully change machine-inferred stance labels without materially degrading semantic integrity.
  • The work positions confidence feedback from stance detectors as a mechanism to systematically escape cocooning effects while keeping intent understandable to humans.

Abstract

Information cocoons on social media limit users' exposure to posts with diverse viewpoints. Modern platforms use stance detection as an important signal in recommendation and ranking pipelines, which can route posts primarily to like-minded audiences and reduce cross-cutting exposure. This restricts the reach of dissenting opinions and hinders constructive discourse. We take the creator's perspective and investigate how content can be revised to reach beyond existing affinity clusters. We present ContentFuzz, a confidence-guided fuzzing framework that rewrites posts while preserving their human-interpreted intent and induces different machine-inferred stance labels. ContentFuzz aims to route posts beyond their original cocoons. Our method guides a large language model (LLM) to generate meaning-preserving rewrites using confidence feedback from stance detection models. Evaluated on four representative stance detection models across three datasets in two languages, ContentFuzz effectively changes machine-classified stance labels, while maintaining semantic integrity with respect to the original content.