52
0

Practical Poisoning Attacks against Retrieval-Augmented Generation

Abstract

Large language models (LLMs) have demonstrated impressive natural language processing abilities but face challenges such as hallucination and outdated knowledge. Retrieval-Augmented Generation (RAG) has emerged as a state-of-the-art approach to mitigate these issues. While RAG enhances LLM outputs, it remains vulnerable to poisoning attacks. Recent studies show that injecting poisoned text into the knowledge database can compromise RAG systems, but most existing attacks assume that the attacker can insert a sufficient number of poisoned texts per query to outnumber correct-answer texts in retrieval, an assumption that is often unrealistic. To address this limitation, we propose CorruptRAG, a practical poisoning attack against RAG systems in which the attacker injects only a single poisoned text, enhancing both feasibility and stealth. Extensive experiments across multiple datasets demonstrate that CorruptRAG achieves higher attack success rates compared to existing baselines.

View on arXiv
@article{zhang2025_2504.03957,
  title={ Practical Poisoning Attacks against Retrieval-Augmented Generation },
  author={ Baolei Zhang and Yuxi Chen and Minghong Fang and Zhuqing Liu and Lihai Nie and Tong Li and Zheli Liu },
  journal={arXiv preprint arXiv:2504.03957},
  year={ 2025 }
}
Comments on this paper