Multiperspectivity as a Resource for Narrative Similarity Prediction

arXiv cs.CL / 3/24/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that narrative similarity prediction is inherently interpretive because multiple equally valid readings of the same text can lead to different similarity judgments, which complicates single-ground-truth semantic benchmarks.
  • It proposes explicitly incorporating this “multiperspectivity” into predictive systems by building an ensemble of 31 LLM personas spanning interpretive-framework practitioners to lay-style characters.
  • Experiments on the SemEval-2026 Task 4 dataset show the approach reaching an accuracy of 0.705, with performance improving as ensemble size increases.
  • The study finds that practitioner personas have lower individual accuracy but make less correlated errors, which drives larger gains under majority voting consistent with Condorcet Jury Theorem-like behavior.
  • Error analysis identifies a negative relationship between gender-focused interpretive vocabulary and accuracy across persona categories, suggesting potential benchmark misalignment or missing interpretations in the ground truth.

Abstract

Predicting narrative similarity can be understood as an inherently interpretive task: different, equally valid readings of the same text can produce divergent interpretations and thus different similarity judgments, posing a fundamental challenge for semantic evaluation benchmarks that encode a single ground truth. Rather than treating this multiperspectivity as a challenge to overcome, we propose to incorporate it in the decision making process of predictive systems. To explore this strategy, we created an ensemble of 31 LLM personas. These range from practitioners following interpretive frameworks to more intuitive, lay-style characters. Our experiments were conducted on the SemEval-2026 Task 4 dataset, where the system achieved an accuracy score of 0.705. Accuracy improves with ensemble size, consistent with Condorcet Jury Theorem-like dynamics under weakened independence. Practitioner personas perform worse individually but produce less correlated errors, yielding larger ensemble gains under majority voting. Our error analysis reveals a consistent negative association between gender-focused interpretive vocabulary and accuracy across all persona categories, suggesting either attention to dimensions not relevant for the benchmark or valid interpretations absent from the ground truth. This finding underscores the need for evaluation frameworks that account for interpretive plurality.