Context-Aware or Context-Insensitive? Assessing LLMs’ Performance in Document-Level Translation

Open Access
Authors
Publication date 2025
Host editors
  • P. Bouillon
  • J. Gerlach
  • S. Girletti
  • L. Volkart
  • R. Rubino
  • R. Sennrich
  • A.C. Farinha
  • M. Gaido
  • J. Daems
  • D. Kenny
  • H. Moniz
  • S. Szoc
Book title MT Summit Geneva 2025 : Machine Translation Summit XX
Book subtitle MTSummit 2025 : June 23-27, 2025, Geneva Switzerland
ISBN (electronic)
  • 9782970189701
Event Machine Translation Summit XX
Volume | Issue number 1
Pages (from-to) 126-137
Number of pages 12
Publisher European Association for Machine Translation
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Large language models (LLMs) are increasingly strong contenders in machine translation. In this work, we focus on document-level translation, where some words cannot be translated without context from outside the sentence. Specifically, we investigate the ability of prominent LLMs to utilize the document context during translation through a perturbation analysis (analyzing models’ robustness to perturbed and randomized document context) and an attribution analysis (examining the contribution of relevant context to the translation). We conduct an extensive evaluation across nine LLMs from diverse model families and training paradigms, including translation-specialized LLMs, alongside two encoder-decoder transformer baselines. We find that LLMs’ improved document-translation performance compared to encoder-decoder models is not reflected in pronoun translation performance. Our analysis highlight the need for context-aware finetuning of LLMs with a focus on relevant parts of the context to improve their reliability for document-level translation.
Document type Conference contribution
Language English
Published at https://aclanthology.org/2025.mtsummit-1.10/
Downloads
2025.mtsummit-1.10 (Final published version)
Permalink to this page
Back