[2603.16430] EngGPT2: Sovereign, Efficient and Open Intelligence
About this article
Abstract page for arXiv paper 2603.16430: EngGPT2: Sovereign, Efficient and Open Intelligence
Computer Science > Computation and Language arXiv:2603.16430 (cs) [Submitted on 17 Mar 2026 (v1), last revised 30 Mar 2026 (this version, v3)] Title:EngGPT2: Sovereign, Efficient and Open Intelligence Authors:G. Ciarfaglia, A. Rosanova, S. Cipolla, J. Bartoli, A. Di Domenico, C. Fioroni, A. Fontana, M. R. Scoleri, M. I. Mone, D. Franchi, M. C. Del Gaudio, A. Leodori, F. Cinti, M. Capozzi, C. Baston, F. Picariello, M. Gabusi, S. Bonura, V. Morreale, I. Bailo View a PDF of the paper titled EngGPT2: Sovereign, Efficient and Open Intelligence, by G. Ciarfaglia and 19 other authors View PDF HTML (experimental) Abstract:EngGPT2-16B-A3B is the latest iteration of Engineering Group's Italian LLM and it's built to be a Sovereign, Efficient and Open model. EngGPT2 is trained on 2.5 trillion tokens - less than Qwen3's 36T or Llama3's 15T - and delivers performance on key benchmarks, including MMLU-Pro, GSM8K, IFEval and HumanEval, comparable to dense models in the 8B-16B range, while requiring one-fifth to half of the inference power, and between one-tenth to one-sixth of the training data and consequent needed training power. Designed as a trained-from-scratch Mixture-of-Experts (MoE) architecture, EngGPT2 features 16 billion parameters with 3 billion active per inference, with expert sizes positioned between those used in GPT-OSS and Qwen3. Approximately 25% of its training corpus consists of Italian-language data, to deliver strong capabilities for European and Italian NLP tasks a...