Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 95% Match Research Paper AI Researchers,LLM Developers,NLP Engineers,Linguists 1 day ago

XIFBench: Evaluating Large Language Models on Multilingual Instruction Following

large-language-models › evaluation
📄 Abstract

Abstract: Large Language Models (LLMs) have demonstrated remarkable instruction-following capabilities across various applications. However, their performance in multilingual settings lacks systematic investigation, with existing evaluations lacking fine-grained constraint analysis across diverse linguistic contexts. We introduce XIFBench, a comprehensive constraint-based benchmark for evaluating multilingual instruction-following abilities of LLMs, comprising 558 instructions with 0-5 additional constraints across five categories (Content, Style, Situation, Format, and Numerical) in six languages spanning different resource levels. To support reliable and consistent cross-lingual evaluation, we implement three methodological innovations: cultural accessibility annotation, constraint-level translation validation, and requirement-based evaluation using English requirements as semantic anchors across languages. Extensive experiments with various LLMs not only quantify performance disparities across resource levels but also provide detailed insights into how language resources, constraint categories, instruction complexity, and cultural specificity influence multilingual instruction-following. Our code and data are available at https://github.com/zhenyuli801/XIFBench.
Authors (8)
Zhenyu Li
Kehai Chen
Yunfei Long
Xuefeng Bai
Yaoyin Zhang
Xuchen Wei
+2 more
Submitted
March 10, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

Introduces XIFBench, a comprehensive constraint-based benchmark for evaluating multilingual instruction-following abilities of LLMs across six languages. It includes methodological innovations like cultural accessibility annotation, constraint-level translation validation, and requirement-based evaluation using English as semantic anchors, quantifying performance disparities.

Business Value

Enables the development of more globally applicable and reliable LLMs, ensuring they can accurately follow instructions and constraints across diverse languages and cultural contexts.