Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Current large language models (LLMs) often suffer from hallucination issues,
i,e, generating content that appears factual but is actually unreliable. A
typical hallucination detection pipeline involves response decomposition (i.e.,
claim extraction), query generation, evidence collection (i.e., search or
retrieval), and claim verification. However, existing methods exhibit
limitations in the first two stages, such as context loss during claim
extraction and low specificity in query generation, resulting in degraded
performance across the hallucination detection pipeline. In this work, we
introduce JointCQ https://github.com/pku0xff/JointCQ, a joint claim-and-query
generation framework designed to construct an effective and efficient
claim-query generator. Our framework leverages elaborately designed evaluation
criteria to filter synthesized training data, and finetunes a language model
for joint claim extraction and query generation, providing reliable and
informative inputs for downstream search and verification. Experimental results
demonstrate that our method outperforms previous methods on multiple
open-domain QA hallucination detection benchmarks, advancing the goal of more
trustworthy and transparent language model systems.
Authors (5)
Fan Xu
Huixuan Zhang
Zhenliang Zhang
Jiahao Wang
Xiaojun Wan
Submitted
October 22, 2025
Key Contributions
Proposes JointCQ, a novel framework for joint claim and query generation to improve factual hallucination detection in LLMs. It addresses limitations in existing methods by preventing context loss during claim extraction and enhancing query specificity, leading to more effective hallucination detection pipelines.
Business Value
Enhances the trustworthiness and reliability of AI-generated content, crucial for applications like news generation, customer service bots, and information retrieval systems, thereby reducing risks associated with misinformation.