Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 85% Match Research Paper NLP Researchers,Machine Learning Engineers,Data Scientists 2 weeks ago

Combining Distantly Supervised Models with In Context Learning for Monolingual and Cross-Lingual Relation Extraction

large-language-models › reasoning
📄 Abstract

Abstract: Distantly Supervised Relation Extraction (DSRE) remains a long-standing challenge in NLP, where models must learn from noisy bag-level annotations while making sentence-level predictions. While existing state-of-the-art (SoTA) DSRE models rely on task-specific training, their integration with in-context learning (ICL) using large language models (LLMs) remains underexplored. A key challenge is that the LLM may not learn relation semantics correctly, due to noisy annotation. In response, we propose HYDRE -- HYbrid Distantly Supervised Relation Extraction framework. It first uses a trained DSRE model to identify the top-k candidate relations for a given test sentence, then uses a novel dynamic exemplar retrieval strategy that extracts reliable, sentence-level exemplars from training data, which are then provided in LLM prompt for outputting the final relation(s). We further extend HYDRE to cross-lingual settings for RE in low-resource languages. Using available English DSRE training data, we evaluate all methods on English as well as a newly curated benchmark covering four diverse low-resource Indic languages -- Oriya, Santali, Manipuri, and Tulu. HYDRE achieves up to 20 F1 point gains in English and, on average, 17 F1 points on Indic languages over prior SoTA DSRE models. Detailed ablations exhibit HYDRE's efficacy compared to other prompting strategies.
Authors (4)
Vipul Rathore
Malik Hammad Faisal
Parag Singla
Mausam
Submitted
October 21, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

This paper introduces HYDRE, a novel framework that combines trained Distantly Supervised Relation Extraction (DSRE) models with In-Context Learning (ICL) using Large Language Models (LLMs). It addresses the challenge of noisy annotations in DSRE by using a dynamic exemplar retrieval strategy to provide reliable sentence-level examples to LLMs, improving relation semantic learning. The framework is also extended to cross-lingual settings for low-resource languages.

Business Value

Improved accuracy in extracting relationships from text can enhance knowledge graph construction, automate information retrieval, and power more sophisticated search engines, leading to better data analysis and decision-making in various industries.