Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
π Abstract
Abstract: Semantic prosody is a collocational meaning formed through the co-occurrence
of a linguistic unit and a consistent series of collocates, which should be
treated separately from semantic meaning. Since words that are literal
translations of each other may have different semantic prosody, more attention
should be paid to this linguistic property to generate accurate translations.
However, current machine translation models cannot handle this problem. To
bridge the gap, we propose an approach to teach machine translation models
about semantic prosody of a specific structure. We focus on Chinese BEI
passives and create a dataset of English-Chinese sentence pairs with the
purpose of demonstrating the negative semantic prosody of BEI passives. Then we
fine-tune OPUS-MT, NLLB-600M and mBART50 models with our dataset for the
English-Chinese translation task. Our results show that fine-tuned MT models
perform better on using BEI passives for translating unfavourable content and
avoid using it for neutral and favourable content. Also, in NLLB-600M, which is
a multilingual model, this knowledge of semantic prosody can be transferred
from English-Chinese translation to other language pairs, such as
Spanish-Chinese.
Authors (4)
Xinyue Ma
Pol Pastells
Mireia FarrΓΊs
Mariona TaulΓ©
Submitted
October 16, 2025
Key Contributions
Addresses the challenge of semantic prosody in machine translation, specifically for English-Chinese passive structures, by creating a specialized dataset and fine-tuning existing MT models. The fine-tuned models demonstrate improved performance in using BEI passives for translating unfavorable content.
Business Value
Enhances the quality and nuance of machine translations, particularly for sensitive content or specific linguistic structures, leading to more reliable cross-cultural communication and better understanding in business and diplomacy.