AMALIA Technical Report: A Fully Open Source Large Language Model for European Portuguese
arXiv cs.CL / 3/30/2026
📰 NewsSignals & Early TrendsModels & Research
Key Points
- The AMALIA technical report introduces a fully open-source large language model designed to better serve European Portuguese (pt-PT), addressing the language’s underrepresentation in training and native evaluations.
- The model is trained with more high-quality pt-PT data during both mid-training and post-training to reduce gaps caused by relying on machine-translated resources.
- The authors release a suite of pt-PT-focused benchmarks, including both translated standard tasks and new datasets aimed at pt-PT generation, linguistic competence, and pt-PT/pt-BR bias measurement.
- Experimental results indicate AMALIA performs comparably to strong baselines on translated benchmarks while delivering substantially improved results on pt-PT-specific evaluations, reinforcing the value of targeted training and native benchmarking.
Related Articles

Black Hat Asia
AI Business

Mr. Chatterbox is a (weak) Victorian-era ethically trained model you can run on your own computer
Simon Willison's Blog
Beyond the Chatbot: Engineering Multi-Agent Ecosystems in 2026
Dev.to

I missed the "fun" part in software development
Dev.to

Hermes Agent: A Self-Improving AI Agent That Runs Anywhere
Dev.to