Teaching LLMs Brazilian Healthcare: Injecting Knowledge from Official Clinical Guidelines
arXiv cs.CL / 5/5/2026
📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research
Key Points
- The paper reports that current LLMs struggle to recall and follow Brazil’s Unified Health System (SUS) guideline knowledge in Brazilian Portuguese, motivating a domain-specific approach.
- It adapts Qwen2.5-14B-Instruct using continual pre-training plus Group Relative Policy Optimization (GRPO) on synthetic data generated from 178 official clinical guidelines (~5.4M tokens).
- The authors introduce HealthBench-BR (1,780 balanced true/false assertions) and PCDT-QA (890 open-ended questions), addressing the lack of Brazilian-protocol-grounded evaluation benchmarks.
- The best 14B-parameter model achieves 83.9% on HealthBench-BR and 85.4% on PCDT-QA, outperforming several larger commercial or web-grounded systems, with ablations highlighting the importance of generator diversity and reinforcement learning.
- All datasets, benchmarks, and model weights are released to enable reproducible clinical NLP research in Brazilian Portuguese, alongside public code and artifacts on GitHub.
Related Articles

Backed by Y Combinator and 20 unicorn founders, Moritz lands $9M
Tech.eu

Why Retail Chargeback Recovery Could Be AgentHansa's First Real PMF
Dev.to

Anthropic Launches AI Services Company with Blackstone & Goldman Sachs
Dev.to

Why B2B Revenue-Recovery Casework Looks Like AgentHansa's Best Early PMF
Dev.to

10 Ways AI Has Become Your Invisible Daily Companion in 2026
Dev.to