Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 91% Match Research Paper NLP Researchers,ML Researchers,Developers of multilingual AI systems 2 weeks ago

Rethinking Cross-lingual Gaps from a Statistical Viewpoint

large-language-models › evaluation
📄 Abstract

Abstract: Any piece of knowledge is usually expressed in one or a handful of natural languages on the web or in any large corpus. Large Language Models (LLMs) act as a bridge by acquiring knowledge from a source language and making it accessible when queried from target languages. Prior research has pointed to a cross-lingual gap, viz., a drop in accuracy when the knowledge is queried in a target language compared to when the query is in the source language. Existing research has rationalized divergence in latent representations in source and target languages as the source of cross-lingual gap. In this work, we take an alternative view and hypothesize that the variance of responses in the target language is the main cause of this gap. For the first time, we formalize the cross-lingual gap in terms of bias-variance decomposition. We present extensive experimental evidence which support proposed formulation and hypothesis. We then reinforce our hypothesis through multiple inference-time interventions that control the variance and reduce the cross-lingual gap. We demonstrate a simple prompt instruction to reduce the response variance, which improved target accuracy by 20-25% across different models.
Authors (5)
Vihari Piratla
Purvam Jain
Darshan Singh
Partha Talukdar
Trevor Cohn
Submitted
October 17, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

This paper re-examines the cross-lingual gap in LLMs from a statistical perspective, hypothesizing that response variance, rather than latent representation divergence, is the main cause. It formalizes the gap using bias-variance decomposition and provides extensive experimental evidence supporting this hypothesis, offering a new framework for understanding and potentially mitigating cross-lingual performance differences.

Business Value

A better statistical understanding of cross-lingual gaps can lead to more equitable and reliable LLM performance across different languages, crucial for global applications and accessibility.