Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

HY-MT1.5 Technical Report

About

In this report, we introduce our latest translation models, HY-MT1.5-1.8B and HY-MT1.5-7B, a new family of machine translation models developed through a holistic training framework tailored for high-performance translation. Our methodology orchestrates a multi-stage pipeline that integrates general and MT-oriented pre-training, supervised fine-tuning, on-policy distillation, and reinforcement learning. HY-MT1.5-1.8B, the 1.8B-parameter model demonstrates remarkable parameter efficiency, comprehensively outperforming significantly larger open-source baselines (e.g., Tower-Plus-72B, Qwen3-32B) and mainstream commercial APIs (e.g., Microsoft Translator, Doubao Translator) in standard Chinese-foreign and English-foreign tasks. It achieves approximately 90% of the performance of ultra-large proprietary models such as Gemini-3.0-Pro, while marginally trailing Gemini-3.0-Pro on WMT25 and Mandarin-minority language benchmarks, it maintains a substantial lead over other competing models. Furthermore, HY-MT1.5-7B establishes a new state-of-the-art for its size class, achieving 95% of Gemini-3.0-Pro's performance on Flores-200 and surpassing it on the challenging WMT25 and Mandarin-minority language test sets. Beyond standard translation, the HY-MT1.5 series supports advanced constraints, including terminology intervention, context-aware translation, and format preservation. Extensive empirical evaluations confirm that both models offer highly competitive, robust solutions for general and specialized translation tasks within their respective parameter scales.

Mao Zheng, Zheng Li, Tao Chen, Mingyang Song, Di Wang• 2025

Related benchmarks

TaskDatasetResultRank
Machine TranslationFLORES+ (test)
spBLEU31.68
128
Machine TranslationWMT24++ v1.0 (test)
XCOMET Score85.77
49
Machine Translation (xx -> zh)FLORES+ latest (test)
spBLEU23.59
30
Machine TranslationWMT 2025 (test)
XCOMET-XXL61.59
17
Machine TranslationFLORES-200 ZH ⇔ XX 2022
XCOMET-XXL0.869
17
Machine TranslationFLORES-200 EN ⇔ XX 2022
XCOMET-XXL90.93
17
Machine TranslationFLORES-200 XX ⇔ XX 2022
XCOMET-XXL80.98
17
Machine TranslationMandarin ⇔ Minority (test)
XCOMET-XXL0.6174
16
Machine TranslationHuman Evaluation ZH⇒EN 2025 (test)
Human Evaluation Score3.01
6
Machine TranslationHuman Evaluation EN⇒ZH 2025 (test)
Human Evaluation Score2.61
6
Showing 10 of 11 rows

Other info

Follow for update