Not All Subjectivity Is the Same! Defining Desiderata for the Evaluation of Subjectivity in NLP

arXiv cs.CL / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that not all forms of subjectivity in NLP are equivalent and proposes seven evaluation desiderata tailored to subjectivity-sensitive models.
  • It frames the desiderata around how subjectivity appears in datasets and how models represent or generate it, with a focus on user-centric outcomes such as visibility of minority perspectives.
  • The authors review the experimental setups of 60 related papers and find several persistent research gaps, including insufficient study of ambiguous versus polyphonic inputs.
  • The review also highlights evaluation shortcomings such as whether subjectivity is actually communicated effectively to users and a lack of consideration for how different desiderata interact with each other.

Abstract

Subjective judgments are part of several NLP datasets and recent work is increasingly prioritizing models whose outputs reflect this diversity of perspectives. Such responses allow us to shed light on minority voices, which are frequently marginalized or obscured by dominant perspectives. It remains a question whether our evaluation practices align with these models' objectives. This position paper proposes seven evaluation desiderata for subjectivity-sensitive models, rooted in how subjectivity is represented in NLP data and models. The desiderata are constructed in a top-down approach, keeping in mind the user-centric impact of such models. We scan the experimental setup of 60 papers and show that various aspects of subjectivity are still understudied: the distinction between ambiguous and polyphonic input, whether subjectivity is effectively expressed to the user, and a lack of interplay between different desiderata, amongst other gaps.