Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 85% Match Research Paper ML Engineers,NLP Researchers,Information Retrieval Specialists,Data Scientists 1 week ago

CustomIR: Unsupervised Fine-Tuning of Dense Embeddings for Known Document Corpora

large-language-models › model-architecture
📄 Abstract

Abstract: Dense embedding models have become critical for modern information retrieval, particularly in RAG pipelines, but their performance often degrades when applied to specialized corpora outside their pre-training distribution. To address thi we introduce \textbf{CustomIR}, a framework for unsupervised adaptation of pre-trained language embedding models to domain-specific corpora using synthetically generated query-document pairs. CustomIR leverages large language models (LLMs) to create diverse queries grounded in a known target corpus, paired with LLM-verified hard negatives, eliminating the need for costly human annotation. Experiments on enterprise email and messaging datasets show that CustomIR consistently improves retrieval effectiveness with small models gaining up to 2.3 points in Recall@10. This performance increase allows these small models to rival the performance of much larger alternatives, allowing for cheaper RAG deployments. These results highlight that targeted synthetic fine-tuning offers a scalable and cost-efficient strategy for increasing domain-specific performance.
Authors (1)
Nathan Paull
Submitted
September 30, 2025
arXiv Category
cs.IR
arXiv PDF

Key Contributions

CustomIR introduces an unsupervised framework to adapt pre-trained dense embedding models to specialized document corpora. By leveraging LLMs for synthetic query generation and hard negative mining, it eliminates the need for human annotation, significantly improving retrieval effectiveness and enabling smaller models to rival larger ones for cost-effective RAG deployments.

Business Value

Enables more cost-effective and performant RAG systems by allowing smaller, cheaper embedding models to achieve high accuracy on domain-specific data, reducing infrastructure costs and improving search relevance for businesses.