Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2407.12784
Cited By
AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases
17 July 2024
Zhaorun Chen
Zhen Xiang
Chaowei Xiao
Dawn Song
Bo Li
LLMAG
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases"
18 / 18 papers shown
Title
POISONCRAFT: Practical Poisoning of Retrieval-Augmented Generation for Large Language Models
Yangguang Shao
Xinjie Lin
Haozheng Luo
Chengshang Hou
G. Xiong
J. Yu
Junzheng Shi
SILM
52
0
0
10 May 2025
AgentXploit: End-to-End Redteaming of Black-Box AI Agents
Zhun Wang
Vincent Siu
Zhe Ye
Tianneng Shi
Yuzhou Nie
Xuandong Zhao
Chenguang Wang
Wenbo Guo
Dawn Song
LLMAG
AAML
36
0
0
09 May 2025
Securing Agentic AI: A Comprehensive Threat Model and Mitigation Framework for Generative AI Agents
Vineeth Sai Narajala
Om Narayan
LLMAG
36
1
0
28 Apr 2025
RAG LLMs are Not Safer: A Safety Analysis of Retrieval-Augmented Generation for Large Language Models
Bang An
Shiyue Zhang
Mark Dredze
61
0
0
25 Apr 2025
Exploiting Fine-Grained Skip Behaviors for Micro-Video Recommendation
Sanghyuck Lee
Sangkeun Park
Jaesung Lee
48
0
0
04 Apr 2025
Emerging Cyber Attack Risks of Medical AI Agents
Jianing Qiu
Lin Li
Jiankai Sun
Hao Wei
Zhe Xu
K. Lam
Wu Yuan
AAML
28
1
0
02 Apr 2025
A Practical Memory Injection Attack against LLM Agents
Shen Dong
Shaocheng Xu
Pengfei He
Y. Li
Jiliang Tang
Tianming Liu
Hui Liu
Zhen Xiang
LLMAG
AAML
43
2
0
05 Mar 2025
Commercial LLM Agents Are Already Vulnerable to Simple Yet Dangerous Attacks
Ang Li
Yin Zhou
Vethavikashini Chithrra Raghuram
Tom Goldstein
Micah Goldblum
AAML
76
7
0
12 Feb 2025
Playing Devil's Advocate: Unmasking Toxicity and Vulnerabilities in Large Vision-Language Models
Abdulkadir Erol
Trilok Padhi
Agnik Saha
Ugur Kursuncu
Mehmet Emin Aktas
45
1
0
17 Jan 2025
Beyond Reward Hacking: Causal Rewards for Large Language Model Alignment
Chaoqi Wang
Zhuokai Zhao
Yibo Jiang
Zhaorun Chen
Chen Zhu
...
Jiayi Liu
Lizhu Zhang
Xiangjun Fan
Hao Ma
Sinong Wang
77
3
0
17 Jan 2025
Agent Security Bench (ASB): Formalizing and Benchmarking Attacks and Defenses in LLM-based Agents
Hanrong Zhang
Jingyuan Huang
Kai Mei
Yifei Yao
Zhenting Wang
Chenlu Zhan
Hongwei Wang
Yongfeng Zhang
AAML
LLMAG
ELM
51
18
0
03 Oct 2024
GuardAgent: Safeguard LLM Agents by a Guard Agent via Knowledge-Enabled Reasoning
Zhen Xiang
Linzhi Zheng
Yanjie Li
Junyuan Hong
Qinbin Li
...
Zidi Xiong
Chulin Xie
Carl Yang
Dawn Song
Bo Li
LLMAG
45
23
0
13 Jun 2024
LLM Agents for Psychology: A Study on Gamified Assessments
Qisen Yang
Zekun Wang
Honghui Chen
Shenzhi Wang
Yifan Pu
Xin Gao
Wenhao Huang
Shiji Song
Gao Huang
LLMAG
127
13
0
19 Feb 2024
COLD-Attack: Jailbreaking LLMs with Stealthiness and Controllability
Xing-ming Guo
Fangxu Yu
Huan Zhang
Lianhui Qin
Bin Hu
AAML
117
69
0
13 Feb 2024
Personalized Autonomous Driving with Large Language Models: Field Experiments
Can Cui
Zichong Yang
Yupeng Zhou
Yunsheng Ma
Juanwu Lu
Lingxi Li
Yaobin Chen
Jitesh Panchal
Ziran Wang
51
17
0
14 Dec 2023
PaperQA: Retrieval-Augmented Generative Agent for Scientific Research
Jakub Lála
Odhran O'Donoghue
Aleksandar Shtedritski
Sam Cox
Samuel G. Rodriques
Andrew D. White
RALM
75
71
0
08 Dec 2023
PoisonPrompt: Backdoor Attack on Prompt-based Large Language Models
Hongwei Yao
Jian Lou
Zhan Qin
SILM
AAML
51
30
0
19 Oct 2023
ReAct: Synergizing Reasoning and Acting in Language Models
Shunyu Yao
Jeffrey Zhao
Dian Yu
Nan Du
Izhak Shafran
Karthik Narasimhan
Yuan Cao
LLMAG
ReLM
LRM
237
2,479
0
06 Oct 2022
1