SG-UniBuc-NLP at SemEval-2026 Task 6: Multi-Head RoBERTa with Chunking for Long-Context Evasion Detection

arXiv cs.AI / 4/30/2026

💬 OpinionModels & Research

Key Points

  • The SG-UniBuc-NLP team presents a system for SemEval-2026 Task 6 (CLARITY) that classifies political interview responses by both coarse-grained clarity (3-way) and fine-grained evasion strategy (9-way).
  • To handle inputs longer than the typical 512-token Transformer limit, they use overlapping sliding-window chunking and aggregate chunk embeddings via element-wise max-pooling.
  • The approach employs a shared RoBERTa-large encoder with two jointly trained, task-specific classification heads under a multi-task objective.
  • For inference, they apply ensembling based on 7-fold stratified cross-validation, achieving Macro-F1 scores of 0.80 (Subtask 1) and 0.51 (Subtask 2), ranking 11th in both subtasks.
  • The results indicate the proposed long-context evasion-detection pipeline is effective for longer political responses, especially for the clarity-focused subtask.

Abstract

We describe our system for SemEval-2026 Task 6 (CLARITY: Unmasking Political Question Evasions), which classifies English political interview responses by coarse-grained clarity (3-way) and fine-grained evasion strategy (9-way). Since responses frequently exceed the 512-token limit of standard Transformer encoders, we apply an overlapping sliding-window chunking strategy with element-wise Max-Pooling aggregation over chunk representations. A shared RoBERTa-large encoder supplies two task-specific heads trained jointly via a multi-task objective, with inference-time ensembling over 7-fold stratified cross-validation. Our system achieves a Macro-F1 of 0.80 on Subtask 1 and 0.51 on Subtask 2, ranking 11th in both subtasks.