Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Lexical Semantic Change Detection (LSCD) is a complex, lemma-level task,
which is usually operationalized based on two subsequently applied usage-level
tasks: First, Word-in-Context (WiC) labels are derived for pairs of usages.
Then, these labels are represented in a graph on which Word Sense Induction
(WSI) is applied to derive sense clusters. Finally, LSCD labels are derived by
comparing sense clusters over time. This modularity is reflected in most LSCD
datasets and models. It also leads to a large heterogeneity in modeling options
and task definitions, which is exacerbated by a variety of dataset versions,
preprocessing options and evaluation metrics. This heterogeneity makes it
difficult to evaluate models under comparable conditions, to choose optimal
model combinations or to reproduce results. Hence, we provide a benchmark
repository standardizing LSCD evaluation. Through transparent implementation
results become easily reproducible and by standardization different components
can be freely combined. The repository reflects the task's modularity by
allowing model evaluation for WiC, WSI and LSCD. This allows for careful
evaluation of increasingly complex model components providing new ways of model
optimization. We use the implemented benchmark to conduct a number of
experiments with recent models and systematically improve the state-of-the-art.
Authors (3)
Dominik Schlechtweg
Sachin Yadav
Nikolay Arefyev
Key Contributions
This paper introduces the LSCD Benchmark, a standardized testbed designed to address the heterogeneity and lack of reproducibility in Lexical Semantic Change Detection research. By providing transparent implementations and consistent evaluation protocols, it enables easier comparison, reproduction, and optimization of LSCD models.
Business Value
Facilitates more reliable research and development in understanding language evolution, which can inform applications in historical text analysis, digital humanities, and potentially improve NLP models' understanding of temporal language shifts.