Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 95% Match Research Paper AI researchers,NLP engineers,Developers of LLM-based applications,Fact-checking organizations 1 week ago

JointCQ: Improving Factual Hallucination Detection with Joint Claim and Query Generation

large-language-models › evaluation
📄 Abstract

Abstract: Current large language models (LLMs) often suffer from hallucination issues, i,e, generating content that appears factual but is actually unreliable. A typical hallucination detection pipeline involves response decomposition (i.e., claim extraction), query generation, evidence collection (i.e., search or retrieval), and claim verification. However, existing methods exhibit limitations in the first two stages, such as context loss during claim extraction and low specificity in query generation, resulting in degraded performance across the hallucination detection pipeline. In this work, we introduce JointCQ https://github.com/pku0xff/JointCQ, a joint claim-and-query generation framework designed to construct an effective and efficient claim-query generator. Our framework leverages elaborately designed evaluation criteria to filter synthesized training data, and finetunes a language model for joint claim extraction and query generation, providing reliable and informative inputs for downstream search and verification. Experimental results demonstrate that our method outperforms previous methods on multiple open-domain QA hallucination detection benchmarks, advancing the goal of more trustworthy and transparent language model systems.
Authors (5)
Fan Xu
Huixuan Zhang
Zhenliang Zhang
Jiahao Wang
Xiaojun Wan
Submitted
October 22, 2025
arXiv Category
cs.CL
arXiv PDF Code

Key Contributions

Proposes JointCQ, a novel framework for joint claim and query generation to improve factual hallucination detection in LLMs. It addresses limitations in existing methods by preventing context loss during claim extraction and enhancing query specificity, leading to more effective hallucination detection pipelines.

Business Value

Enhances the trustworthiness and reliability of AI-generated content, crucial for applications like news generation, customer service bots, and information retrieval systems, thereby reducing risks associated with misinformation.

View Code on GitHub