Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

EngGPT2: Sovereign, Efficient and Open Intelligence

About

EngGPT2-16B-A3B is the latest iteration of Engineering Group's Italian LLM and it's built to be a Sovereign, Efficient and Open model. EngGPT2 is trained on 2.5 trillion tokens - less than Qwen3's 36T or Llama3's 15T - and delivers performance on key benchmarks, including MMLU-Pro, GSM8K, IFEval and HumanEval, comparable to dense models in the 8B-16B range, while requiring one-fifth to half of the inference power, and between one-tenth to one-sixth of the training data and consequent needed training power. Designed as a trained-from-scratch Mixture-of-Experts (MoE) architecture, EngGPT2 features 16 billion parameters with 3 billion active per inference, with expert sizes positioned between those used in GPT-OSS and Qwen3. Approximately 25% of its training corpus consists of Italian-language data, to deliver strong capabilities for European and Italian NLP tasks among models of similar scale. This efficiency aims to position EngGPT2 as a key contributor to the growing portfolio of open-weight European models, combining performance and efficiency with full alignment to the EU AI Act. EngGPT2 is also a single model capable of multiple reasoning modes: non-reasoning, reasoning in Italian or English, and turbo-reasoning (a concise, bullet-point style reasoning available in both languages designed for real-time reasoning use cases). EngGPT2 aims to set a new standard for resource-conscious, high-performance LLMs tailored to European and Italian contexts.

G. Ciarfaglia, A. Rosanova, S. Cipolla, J. Bartoli, A. Di Domenico, C. Fioroni, A. Fontana, M. R. Scoleri, M. I. Mone, D. Franchi, M. C. Del Gaudio, A. Leodori, F. Cinti, M. Capozzi, C. Baston, F. Picariello, M. Gabusi, S. Bonura, V. Morreale, I. Bailo• 2026

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval
Pass@164
1036
Instruction FollowingIFEval
IFEval Accuracy72
625
Multi-task Language UnderstandingMMLU-Redux
Accuracy75.5
44
Knowledge ReasoningMMLU-Pro
MMLU-Pro Knowledge Reasoning Score57.3
40
Massive Multitask Language UnderstandingMMLU-Pro
Accuracy (MMLU-Pro)57.3
38
Mathematical ReasoningAIME 26
Score70
28
Multi-task Language UnderstandingMMLU-IT
Accuracy65.5
16
Question AnsweringARC-IT
Accuracy85.6
16
Function CallingBFCL
Accuracy48.5
14
Knowledge & ReasoningMMLU-Redux
MMLU Redux Score75.5
8
Showing 10 of 10 rows

Other info

Follow for update