Automatic Machine Translation Evaluation in Many Languages via Zero-Shot Paraphrasing
About
We frame the task of machine translation evaluation as one of scoring machine translation output with a sequence-to-sequence paraphraser, conditioned on a human reference. We propose training the paraphraser as a multilingual NMT system, treating paraphrasing as a zero-shot translation task (e.g., Czech to Czech). This results in the paraphraser's output mode being centered around a copy of the input sequence, which represents the best case scenario where the MT system output matches a human reference. Our method is simple and intuitive, and does not require human judgements for training. Our single model (trained in 39 languages) outperforms or statistically ties with all prior metrics on the WMT 2019 segment-level shared metrics task in all languages (excluding Gujarati where the model had no training data). We also explore using our model for the task of quality estimation as a metric--conditioning on the source instead of the reference--and find that it significantly outperforms every submission to the WMT 2019 shared task on quality estimation in every language pair.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Summarization Evaluation | SummEval | Coherence23.3 | 41 | |
| Machine Translation Meta-evaluation | WMT Metrics Shared Task Segment-level 2023 (Primary submissions) | Avg Correlation0.593 | 33 | |
| Dialogue Evaluation Human Correlation | Topical-Chat | Naturalness Pearson (r)0.04 | 26 | |
| Data-to-text evaluation | SFHOT | Spearman Correlation0.196 | 24 | |
| Data-to-text evaluation | SFRES | Spearman Correlation0.155 | 24 | |
| Factuality Evaluation | QAGS XSUM | Pearson Correlation0.025 | 19 | |
| Factuality Evaluation | Rank19 | Accuracy78 | 13 | |
| Dialect Robustness | PT | Success Rate53 | 11 | |
| Dialect Robustness | ZH | Success Rate47 | 11 | |
| Dialect Robustness | EN | Success Rate51 | 11 |