Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 90% Match Research Paper NLP researchers,Machine learning engineers,Computational linguists 3 weeks ago

Multilinguality Does not Make Sense: Investigating Factors Behind Zero-Shot Transfer in Sense-Aware Tasks

large-language-models › evaluation
📄 Abstract

Abstract: Cross-lingual transfer is central to modern NLP, enabling models to perform tasks in languages different from those they were trained on. A common assumption is that training on more languages improves zero-shot transfer. We test this on sense-aware tasks-polysemy and lexical semantic change-and find that multilinguality is not necessary for effective transfer. Our large-scale analysis across 28 languages reveals that other factors, such as differences in pretraining and fine-tuning data and evaluation artifacts, better explain the perceived benefits of multilinguality. We also release fine-tuned models and provide empirical baselines to support future research. While focused on two sense-aware tasks, our findings offer broader insights into cross-lingual transfer, especially for low-resource languages.
Authors (2)
Roksana Goworek
Haim Dubossarsky
Submitted
May 30, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

Challenges the assumption that multilinguality is necessary for effective zero-shot transfer in sense-aware tasks. Through large-scale analysis across 28 languages, it identifies pretraining/fine-tuning data and evaluation artifacts as more significant factors, offering a refined understanding of cross-lingual transfer, especially for low-resource languages.

Business Value

Optimizes the development of multilingual NLP models by focusing on more impactful factors than just language count, potentially reducing training costs and improving performance for low-resource languages.