SG-UniBuc-NLP at SemEval-2026 Task 6: Multi-Head RoBERTa with Chunking for Long-Context Evasion Detection
arXiv cs.AI / 4/30/2026
💬 OpinionModels & Research
Key Points
- The SG-UniBuc-NLP team presents a system for SemEval-2026 Task 6 (CLARITY) that classifies political interview responses by both coarse-grained clarity (3-way) and fine-grained evasion strategy (9-way).
- To handle inputs longer than the typical 512-token Transformer limit, they use overlapping sliding-window chunking and aggregate chunk embeddings via element-wise max-pooling.
- The approach employs a shared RoBERTa-large encoder with two jointly trained, task-specific classification heads under a multi-task objective.
- For inference, they apply ensembling based on 7-fold stratified cross-validation, achieving Macro-F1 scores of 0.80 (Subtask 1) and 0.51 (Subtask 2), ranking 11th in both subtasks.
- The results indicate the proposed long-context evasion-detection pipeline is effective for longer political responses, especially for the clarity-focused subtask.
Related Articles
Vector DB and ANN vs PHE conflict, is there a practical workaround? [D]
Reddit r/MachineLearning

Azure Weekly: Microsoft and OpenAI Restructure Partnership as GPT-5.5 Lands in Foundry
Dev.to

Vibe coding is a tool, not a shortcut. Most people are using it wrong.
Dev.to

Automating YouTube Content Creation with Artificial Intelligence
Dev.to

Memento: Fine-tuning LLM Agents without Fine-tuning LLMs
Dev.to