MATH-PT: A Math Reasoning Benchmark for European and Brazilian Portuguese
arXiv cs.CL / 4/30/2026
📰 NewsSignals & Early TrendsModels & Research
Key Points
- The paper introduces Math-PT, a new Portuguese (European and Brazilian) math reasoning benchmark containing 1,729 problems sourced from native Portuguese materials such as olympiads, competitions, and exams.
- It argues that existing math-reasoning evaluations are heavily linguistically biased toward English (or English translations), limiting fairness and usefulness across languages.
- The authors evaluate current state-of-the-art LLMs on Math-PT and find that frontier reasoning models perform better on multiple-choice questions than open-weight models.
- The study also shows a drop in performance for questions that include figures and for open-ended questions, highlighting ongoing weaknesses in multimodal and free-form reasoning.
- To support further work, the benchmark dataset and the model outputs are released for public use.
Related Articles
Vector DB and ANN vs PHE conflict, is there a practical workaround? [D]
Reddit r/MachineLearning

Azure Weekly: Microsoft and OpenAI Restructure Partnership as GPT-5.5 Lands in Foundry
Dev.to

Proven Patterns for OpenAI Codex in 2026: Prompts, Validation, and Gateway Governance
Dev.to

Vibe coding is a tool, not a shortcut. Most people are using it wrong.
Dev.to

How VS Code v1.117.0 Changes Collaboration with GitHub Copilot as Co-Author
Dev.to