AI Navigate

EngGPT2: Sovereign, Efficient and Open Intelligence

arXiv cs.CL / 3/18/2026

📰 NewsIdeas & Deep AnalysisIndustry & Market MovesModels & Research

Key Points

  • EngGPT2-16B-A3B is a newly announced Italian LLM from Engineering Group designed to be sovereign, efficient, and open, with explicit alignment to the EU AI Act.
  • It uses a trained-from-scratch Mixture-of-Experts architecture with 16 billion parameters and about 3 billion active per inference, delivering competitive performance on benchmarks such as MMLU-Pro, GSM8K, IFEval and HumanEval relative to 8-16B dense models.
  • It was trained on 2.5 trillion tokens, with roughly 25% Italian-language data to strengthen European and Italian NLP capabilities at this scale.
  • EngGPT2 claims substantial efficiency improvements, requiring one-fifth to one-half of the inference power and one-tenth to one-sixth of the training data and training power of comparable models.
  • It supports multiple reasoning modes, including non-reasoning, Italian-English reasoning, and turbo-reasoning, positioning it for real-time, multilingual use cases and as a model in open-weight European AI ecosystems.

Abstract

EngGPT2-16B-A3B is the latest iteration of Engineering Group's Italian LLM and it's built to be a Sovereign, Efficient and Open model. EngGPT2 is trained on 2.5 trillion tokens - less than Qwen3's 36T or Llama3's 15T - and delivers performance on key benchmarks, including MMLU-Pro, GSM8K, IFEval and HumanEval, comparable to dense models in the 8B-16B range, while requiring one-fifth to half of the inference power, and between one-tenth to one-sixth of the training data and consequent needed training power. Designed as a trained-from-scratch Mixture-of-Experts (MoE) architecture, EngGPT2 features 16 billion parameters with 3 billion active per inference, with expert sizes positioned between those used in GPT-OSS and Qwen3. Approximately 25% of its training corpus consists of Italian-language data, to deliver strong capabilities for European and Italian NLP tasks among models of similar scale. This efficiency aims to position EngGPT2 as a key contributor to the growing portfolio of open-weight European models, combining performance and efficiency with full alignment to the EU AI Act. EngGPT2 is also a single model capable of multiple reasoning modes: non-reasoning, reasoning in Italian or English, and turbo-reasoning (a concise, bullet-point style reasoning available in both languages designed for real-time reasoning use cases). EngGPT2 aims to set a new standard for resource-conscious, high-performance LLMs tailored to European and Italian contexts.