Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Error Analysis Prompting Enables Human-Like Translation Evaluation in Large Language Models

About

Generative large language models (LLMs), e.g., ChatGPT, have demonstrated remarkable proficiency across several NLP tasks, such as machine translation, text summarization. Recent research (Kocmi and Federmann, 2023) has shown that utilizing LLMs for assessing the quality of machine translation (MT) achieves state-of-the-art performance at the system level but \textit{performs poorly at the segment level}. To further improve the performance of LLMs on MT quality assessment, we investigate several prompting designs, and propose a new prompting method called \textbf{\texttt{Error Analysis Prompting}} (EAPrompt) by combining Chain-of-Thoughts (Wei et al., 2022) and Error Analysis (Lu et al., 2023). This technique emulates the commonly accepted human evaluation framework - Multidimensional Quality Metrics (MQM, Freitag et al. (2021)) and \textit{produces explainable and reliable MT evaluations at both the system and segment level}. Experimental Results from the WMT22 metrics shared task validate the effectiveness of EAPrompt on various LLMs, with different structures. Further analysis confirms that EAPrompt effectively distinguishes major errors from minor ones, while also sharing a similar distribution of the number of errors with MQM. These findings highlight the potential of EAPrompt as a human-like evaluator prompting technique for MT evaluation.

Qingyu Lu, Baopu Qiu, Liang Ding, Kanjian Zhang, Tom Kocmi, Dacheng Tao• 2023

Related benchmarks

TaskDatasetResultRank
Machine Translation Evaluation MetricWMT MQM 23
Acc90.3
27
Machine Translation EvaluationWMT MQM System-level 22
Overall Score91.2
19
Machine Translation EvaluationWMT MQM Segment-level 22
Score (En-De)56.7
19
Machine Translation EvaluationWMT MQM 2022 (test)
Accuracy (System, 3 LPs)91.2
16
Machine Translation Meta-evaluationWMT Zh-En (subset of 600 samples) 2022
Kendall Correlation0.4597
2
Showing 5 of 5 rows

Other info

Code

Follow for update