Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 90% Match Research Paper Computational Linguists,NLP Researchers,Digital Humanities Scholars,Historians 1 week ago

The LSCD Benchmark: a Testbed for Diachronic Word Meaning Tasks

large-language-models › evaluation
📄 Abstract

Abstract: Lexical Semantic Change Detection (LSCD) is a complex, lemma-level task, which is usually operationalized based on two subsequently applied usage-level tasks: First, Word-in-Context (WiC) labels are derived for pairs of usages. Then, these labels are represented in a graph on which Word Sense Induction (WSI) is applied to derive sense clusters. Finally, LSCD labels are derived by comparing sense clusters over time. This modularity is reflected in most LSCD datasets and models. It also leads to a large heterogeneity in modeling options and task definitions, which is exacerbated by a variety of dataset versions, preprocessing options and evaluation metrics. This heterogeneity makes it difficult to evaluate models under comparable conditions, to choose optimal model combinations or to reproduce results. Hence, we provide a benchmark repository standardizing LSCD evaluation. Through transparent implementation results become easily reproducible and by standardization different components can be freely combined. The repository reflects the task's modularity by allowing model evaluation for WiC, WSI and LSCD. This allows for careful evaluation of increasingly complex model components providing new ways of model optimization. We use the implemented benchmark to conduct a number of experiments with recent models and systematically improve the state-of-the-art.
Authors (3)
Dominik Schlechtweg
Sachin Yadav
Nikolay Arefyev
Submitted
March 29, 2024
arXiv Category
cs.CL
arXiv PDF

Key Contributions

This paper introduces the LSCD Benchmark, a standardized testbed designed to address the heterogeneity and lack of reproducibility in Lexical Semantic Change Detection research. By providing transparent implementations and consistent evaluation protocols, it enables easier comparison, reproduction, and optimization of LSCD models.

Business Value

Facilitates more reliable research and development in understanding language evolution, which can inform applications in historical text analysis, digital humanities, and potentially improve NLP models' understanding of temporal language shifts.