Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Large language models (LLMs) have transformed natural language processing
(NLP), enabling applications from content generation to decision support.
Retrieval-Augmented Generation (RAG) improves LLMs by incorporating external
knowledge but also introduces security risks, particularly from data poisoning,
where the attacker injects poisoned texts into the knowledge database to
manipulate system outputs. While various defenses have been proposed, they
often struggle against advanced attacks. To address this, we introduce RAGuard,
a detection framework designed to identify poisoned texts. RAGuard first
expands the retrieval scope to increase the proportion of clean texts, reducing
the likelihood of retrieving poisoned content. It then applies chunk-wise
perplexity filtering to detect abnormal variations and text similarity
filtering to flag highly similar texts. This non-parametric approach enhances
RAG security, and experiments on large-scale datasets demonstrate its
effectiveness in detecting and mitigating poisoning attacks, including strong
adaptive attacks.
Authors (7)
Zirui Cheng
Jikai Sun
Anjun Gao
Yueyang Quan
Zhuqing Liu
Xiaohua Hu
+1 more
Submitted
October 28, 2025
Key Contributions
Introduces RAGuard, a detection framework to identify poisoned texts in Retrieval-Augmented Generation (RAG) systems. It uses expanded retrieval scope, chunk-wise perplexity filtering, and text similarity filtering to enhance RAG security against advanced poisoning attacks.
Business Value
Protects businesses relying on LLMs for critical applications (e.g., customer support, content creation) from malicious manipulation of their knowledge bases, ensuring reliable and trustworthy outputs.