Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 90% Match Research Paper NLP Researchers,Machine Learning Engineers,Computational Linguists 1 week ago

Dependency Parsing is More Parameter-Efficient with Normalization

large-language-models β€Ί model-architecture
πŸ“„ Abstract

Abstract: Dependency parsing is the task of inferring natural language structure, often approached by modeling word interactions via attention through biaffine scoring. This mechanism works like self-attention in Transformers, where scores are calculated for every pair of words in a sentence. However, unlike Transformer attention, biaffine scoring does not use normalization prior to taking the softmax of the scores. In this paper, we provide theoretical evidence and empirical results revealing that a lack of normalization necessarily results in overparameterized parser models, where the extra parameters compensate for the sharp softmax outputs produced by high variance inputs to the biaffine scoring function. We argue that biaffine scoring can be made substantially more efficient by performing score normalization. We conduct experiments on semantic and syntactic dependency parsing in multiple languages, along with latent graph inference on non-linguistic data, using various settings of a $k$-hop parser. We train $N$-layer stacked BiLSTMs and evaluate the parser's performance with and without normalizing biaffine scores. Normalizing allows us to achieve state-of-the-art performance with fewer samples and trainable parameters. Code: https://github.com/paolo-gajo/EfficientSDP
Authors (4)
Paolo Gajo
Domenic Rosati
Hassan Sajjad
Alberto BarrΓ³n-CedeΓ±o
Submitted
May 26, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

This paper demonstrates that the lack of normalization in biaffine scoring for dependency parsing leads to overparameterized models. It provides theoretical evidence and empirical results showing that score normalization can significantly improve parameter efficiency. This work offers a practical method to make parsers more efficient without sacrificing accuracy.

Business Value

More efficient NLP models for tasks like parsing can reduce computational costs and latency, making advanced language understanding more accessible for real-time applications.