STaR-DRO: Stateful Tsallis Reweighting for Group-Robust Structured Prediction
arXiv cs.LG / 4/14/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a two-part framework for structured prediction under ambiguity, label skew, and group heterogeneity, combining task-agnostic prompting (XML-structured instructions, disambiguation/verification, schema constraints, and self-validation) with robust fine-tuning.
- It introduces STaR-DRO, a stateful robust optimization method that uses Tsallis mirror descent with momentum-smoothed, centered group-loss signals and bounded excess-only reweighting to upweight only persistently hard groups.
- The approach is evaluated on EPPC Miner, a benchmark for extracting hierarchical labels and evidence spans from secure patient-provider messages, targeting both correctness and evidence groundedness.
- Results show prompting improves zero-shot structured extraction by +15.44 average F1 across multiple Llama models, and adding STaR-DRO to supervised fine-tuning improves hardest semantic decisions (e.g., Llama-3.3-70B-Instruct Code F1 79.24→81.47; Sub-code 67.78→69.30) while preserving span performance and reducing group-wise validation cross-entropy by up to 29.6%.
- The authors argue the gains matter for real-world communication mining reliability in patient-centered care due to better handling of rare and clinically consequential categories.
Related Articles

Black Hat Asia
AI Business

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Don't forget, there is more than forgetting: new metrics for Continual Learning
Dev.to

Microsoft MAI-Image-2-Efficient Review 2026: The AI Image Model Built for Production Scale
Dev.to
Bit of a strange question?
Reddit r/artificial