38

"Someone Hid It": Query-Agnostic Black-Box Attacks on LLM-Based Retrieval

Jiate Li
Defu Cao
Li Li
Wei Yang
Yuehan Qin
Chenxiao Yu
Tiannuo Yang
Ryan A. Rossi
Yan Liu
Xiyang Hu
Yue Zhao
Main:8 Pages
7 Figures
Bibliography:3 Pages
6 Tables
Appendix:3 Pages
Abstract

Large language models (LLMs) have been serving as effective backbones for retrieval systems, including Retrieval-Augmentation-Generation (RAG), Dense Information Retriever (IR), and Agent Memory Retrieval. Recent studies have demonstrated that such LLM-based Retrieval (LLMR) is vulnerable to adversarial attacks, which manipulates documents by token-level injections and enables adversaries to either boost or diminish these documents in retrieval tasks. However, existing attack studies mainly (1) presume a known query is given to the attacker, and (2) highly rely on access to the victim model's parameters or interactions, which are hardly accessible in real-world scenarios, leading to limited validity.

View on arXiv
Comments on this paper