Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Multilingual generation with large language models (LLMs) is often of poor
quality for mid- to low-resource languages, but the causes for this are not
well-understood. We first demonstrate the existence of an implicit
task-solving-->translation pipeline for generation, whereby the model first
solves the required task in a largely target-language-agnostic manner, and
subsequently translates answer concepts into the intended target language. We
hypothesize that the failure of the translation stage, despite task-solving
success, is an important culprit for the observed low quality of final outputs,
and formalize this as the translation barrier hypothesis. We quantify the
extent to which either stage in the pipeline is responsible for final failure
for a word translation task across 108 language pairs, and find that the
translation barrier explains a dominant portion of error for a majority of
language pairs, and is especially severe for low-resource target languages. Our
results highlight an important bottleneck for end-to-end multilingual
generation, relevant for future work seeking to improve multilinguality in
LLMs.
Authors (7)
Niyati Bafna
Tianjian Li
Kenton Murray
David R. Mortensen
David Yarowsky
Hale Sirin
+1 more
Key Contributions
Proposes the 'Translation Barrier Hypothesis', suggesting that poor multilingual generation in LLMs stems from an implicit task-solving -> translation pipeline where the translation stage fails, especially for low-resource languages. Quantifies this barrier across 108 language pairs, showing it's a dominant error source.
Business Value
Helps developers and researchers identify key bottlenecks in multilingual AI systems, leading to more effective strategies for improving global language support and reducing development costs.