Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Chain-of-Thought Reasoning Improves Context-Aware Translation with Large Language Models

About

This paper assesses the ability of large language models (LLMs) to translate texts that include inter-sentential dependencies. We use the English-French DiscEvalMT benchmark (Bawden et al., 2018) with pairs of sentences containing translation challenges for pronominal anaphora and lexical cohesion. We evaluate 12 LLMs from the DeepSeek-R1, GPT, Llama, Mistral and Phi families on two tasks: (1) distinguish a correct translation from a wrong but plausible one; and (2) generate a correct translation. We compare prompts that encourage chain-of-thought reasoning with those that do not. The best models take advantage of reasoning and reach about 90% accuracy on the first task and COMET scores of about 92% on the second task, with GPT-4, GPT-4o and Phi standing out. Moreover, we observe a "wise get wiser" effect: the improvements through reasoning are larger for models that already perform well without reasoning.

Shabnam Ataee, Hugo Huart, Andrei Popescu-Belis• 2025

Related benchmarks

TaskDatasetResultRank
Contextual Machine Translationanaphora benchmark
BLEU0.54
25
Machine Translationlexical choice benchmark
BLEU54
25
Machine Translationanaphora benchmark (val+test)
BLEU54.41
24
Translationlexical choice (val+test)
BLEU54.15
24
Showing 4 of 4 rows

Other info

Follow for update