Judge Like Human Examiners: A Weighted Importance Multi-Point Evaluation Framework for Generative Tasks with Long-form Answers

arXiv cs.CL / 4/14/2026

💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper addresses the difficulty of evaluating long-form generative responses, arguing that reference answers contain multiple complementary factors that should be separated for detailed scoring.
  • It proposes the Weighted Importance Multi-Point Evaluation (WIMPE) framework, which decomposes reference answers into weighted, context-bound scoring points to support fine-grained assessment.
  • Two metrics—Weighted Point-wise Alignment (WPA) and Point-wise Conflict Penalty (PCP)—are introduced to measure how well a model’s response aligns with reference points and how much it contradicts them.
  • Experiments across 10 generative tasks reportedly show that WIMPE correlates better with human annotations than prior rubric- or checklist-based approaches.

Abstract

Evaluating the quality of model responses remains challenging in generative tasks with long-form answers, as the expected answers usually contain multiple semantically distinct yet complementary factors that should be factorized for fine-grained assessment. Recent evaluation methods resort to relying on either task-level rubrics or question-aware checklists. However, they still 1) struggle to assess whether a response is genuinely grounded in provided contexts; 2) fail to capture the heterogeneous importance of different aspects of reference answers. Inspired by human examiners, we propose a Weighted Importance Multi-Point Evaluation (WIMPE) framework, which factorizes each reference answer into weighted context-bound scoring points. Two complementary metrics, namely Weighted Point-wise Alignment (WPA) and Point-wise Conflict Penalty (PCP), are designed to measure the alignment and contradiction between model responses and reference answers. Extensive experiments on 10 generative tasks demonstrate that WIMPE achieves higher correlations with human annotations.