Context-Aware or Context-Insensitive? Assessing LLMs’ Performance in Document-Level Translation
| Authors | |
|---|---|
| Publication date | 2025 |
| Host editors |
|
| Book title | MT Summit Geneva 2025 : Machine Translation Summit XX |
| Book subtitle | MTSummit 2025 : June 23-27, 2025, Geneva Switzerland |
| ISBN (electronic) |
|
| Event | Machine Translation Summit XX |
| Volume | Issue number | 1 |
| Pages (from-to) | 126-137 |
| Number of pages | 12 |
| Publisher | European Association for Machine Translation |
| Organisations |
|
| Abstract |
Large language models (LLMs) are increasingly strong contenders in machine translation. In this work, we focus on document-level translation, where some words cannot be translated without context from outside the sentence. Specifically, we investigate the ability of prominent LLMs to utilize the document context during translation through a perturbation analysis (analyzing models’ robustness to perturbed and randomized document context) and an attribution analysis (examining the contribution of relevant context to the translation). We conduct an extensive evaluation across nine LLMs from diverse model families and training paradigms, including translation-specialized LLMs, alongside two encoder-decoder transformer baselines. We find that LLMs’ improved document-translation performance compared to encoder-decoder models is not reflected in pronoun translation performance. Our analysis highlight the need for context-aware finetuning of LLMs with a focus on relevant parts of the context to improve their reliability for document-level translation.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://aclanthology.org/2025.mtsummit-1.10/ |
| Downloads |
2025.mtsummit-1.10
(Final published version)
|
| Permalink to this page | |