ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2407.12784
  4. Cited By
AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge
  Bases

AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases

17 July 2024
Zhaorun Chen
Zhen Xiang
Chaowei Xiao
Dawn Song
Bo Li
    LLMAG
    AAML
ArXivPDFHTML

Papers citing "AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases"

18 / 18 papers shown
Title
POISONCRAFT: Practical Poisoning of Retrieval-Augmented Generation for Large Language Models
POISONCRAFT: Practical Poisoning of Retrieval-Augmented Generation for Large Language Models
Yangguang Shao
Xinjie Lin
Haozheng Luo
Chengshang Hou
G. Xiong
J. Yu
Junzheng Shi
SILM
52
0
0
10 May 2025
AgentXploit: End-to-End Redteaming of Black-Box AI Agents
AgentXploit: End-to-End Redteaming of Black-Box AI Agents
Zhun Wang
Vincent Siu
Zhe Ye
Tianneng Shi
Yuzhou Nie
Xuandong Zhao
Chenguang Wang
Wenbo Guo
Dawn Song
LLMAG
AAML
36
0
0
09 May 2025
Securing Agentic AI: A Comprehensive Threat Model and Mitigation Framework for Generative AI Agents
Securing Agentic AI: A Comprehensive Threat Model and Mitigation Framework for Generative AI Agents
Vineeth Sai Narajala
Om Narayan
LLMAG
36
1
0
28 Apr 2025
RAG LLMs are Not Safer: A Safety Analysis of Retrieval-Augmented Generation for Large Language Models
RAG LLMs are Not Safer: A Safety Analysis of Retrieval-Augmented Generation for Large Language Models
Bang An
Shiyue Zhang
Mark Dredze
61
0
0
25 Apr 2025
Exploiting Fine-Grained Skip Behaviors for Micro-Video Recommendation
Exploiting Fine-Grained Skip Behaviors for Micro-Video Recommendation
Sanghyuck Lee
Sangkeun Park
Jaesung Lee
48
0
0
04 Apr 2025
Emerging Cyber Attack Risks of Medical AI Agents
Emerging Cyber Attack Risks of Medical AI Agents
Jianing Qiu
Lin Li
Jiankai Sun
Hao Wei
Zhe Xu
K. Lam
Wu Yuan
AAML
28
1
0
02 Apr 2025
A Practical Memory Injection Attack against LLM Agents
Shen Dong
Shaocheng Xu
Pengfei He
Y. Li
Jiliang Tang
Tianming Liu
Hui Liu
Zhen Xiang
LLMAG
AAML
43
2
0
05 Mar 2025
Commercial LLM Agents Are Already Vulnerable to Simple Yet Dangerous Attacks
Commercial LLM Agents Are Already Vulnerable to Simple Yet Dangerous Attacks
Ang Li
Yin Zhou
Vethavikashini Chithrra Raghuram
Tom Goldstein
Micah Goldblum
AAML
76
7
0
12 Feb 2025
Playing Devil's Advocate: Unmasking Toxicity and Vulnerabilities in Large Vision-Language Models
Playing Devil's Advocate: Unmasking Toxicity and Vulnerabilities in Large Vision-Language Models
Abdulkadir Erol
Trilok Padhi
Agnik Saha
Ugur Kursuncu
Mehmet Emin Aktas
45
1
0
17 Jan 2025
Beyond Reward Hacking: Causal Rewards for Large Language Model Alignment
Beyond Reward Hacking: Causal Rewards for Large Language Model Alignment
Chaoqi Wang
Zhuokai Zhao
Yibo Jiang
Zhaorun Chen
Chen Zhu
...
Jiayi Liu
Lizhu Zhang
Xiangjun Fan
Hao Ma
Sinong Wang
77
3
0
17 Jan 2025
Agent Security Bench (ASB): Formalizing and Benchmarking Attacks and Defenses in LLM-based Agents
Agent Security Bench (ASB): Formalizing and Benchmarking Attacks and Defenses in LLM-based Agents
Hanrong Zhang
Jingyuan Huang
Kai Mei
Yifei Yao
Zhenting Wang
Chenlu Zhan
Hongwei Wang
Yongfeng Zhang
AAML
LLMAG
ELM
51
18
0
03 Oct 2024
GuardAgent: Safeguard LLM Agents by a Guard Agent via Knowledge-Enabled Reasoning
GuardAgent: Safeguard LLM Agents by a Guard Agent via Knowledge-Enabled Reasoning
Zhen Xiang
Linzhi Zheng
Yanjie Li
Junyuan Hong
Qinbin Li
...
Zidi Xiong
Chulin Xie
Carl Yang
Dawn Song
Bo Li
LLMAG
45
23
0
13 Jun 2024
LLM Agents for Psychology: A Study on Gamified Assessments
LLM Agents for Psychology: A Study on Gamified Assessments
Qisen Yang
Zekun Wang
Honghui Chen
Shenzhi Wang
Yifan Pu
Xin Gao
Wenhao Huang
Shiji Song
Gao Huang
LLMAG
127
13
0
19 Feb 2024
COLD-Attack: Jailbreaking LLMs with Stealthiness and Controllability
COLD-Attack: Jailbreaking LLMs with Stealthiness and Controllability
Xing-ming Guo
Fangxu Yu
Huan Zhang
Lianhui Qin
Bin Hu
AAML
117
69
0
13 Feb 2024
Personalized Autonomous Driving with Large Language Models: Field
  Experiments
Personalized Autonomous Driving with Large Language Models: Field Experiments
Can Cui
Zichong Yang
Yupeng Zhou
Yunsheng Ma
Juanwu Lu
Lingxi Li
Yaobin Chen
Jitesh Panchal
Ziran Wang
51
17
0
14 Dec 2023
PaperQA: Retrieval-Augmented Generative Agent for Scientific Research
PaperQA: Retrieval-Augmented Generative Agent for Scientific Research
Jakub Lála
Odhran O'Donoghue
Aleksandar Shtedritski
Sam Cox
Samuel G. Rodriques
Andrew D. White
RALM
75
71
0
08 Dec 2023
PoisonPrompt: Backdoor Attack on Prompt-based Large Language Models
PoisonPrompt: Backdoor Attack on Prompt-based Large Language Models
Hongwei Yao
Jian Lou
Zhan Qin
SILM
AAML
51
30
0
19 Oct 2023
ReAct: Synergizing Reasoning and Acting in Language Models
ReAct: Synergizing Reasoning and Acting in Language Models
Shunyu Yao
Jeffrey Zhao
Dian Yu
Nan Du
Izhak Shafran
Karthik Narasimhan
Yuan Cao
LLMAG
ReLM
LRM
237
2,479
0
06 Oct 2022
1