Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
All Papers
0 / 0 papers shown
Title
Home
Papers
2507.02735
Cited By
v1
v2 (latest)
Meta SecAlign: A Secure Foundation LLM Against Prompt Injection Attacks
3 July 2025
Sizhe Chen
Arman Zharmagambetov
David Wagner
Chuan Guo
AAML
Re-assign community
ArXiv (abs)
PDF
HTML
Github (1579★)
Papers citing
"Meta SecAlign: A Secure Foundation LLM Against Prompt Injection Attacks"
17 / 17 papers shown
Title
BrowseSafe: Understanding and Preventing Prompt Injection Within AI Browser Agents
Kaiyuan Zhang
Mark Tenenholtz
Kyle Polley
Jerry Ma
Denis Yarats
Ninghui Li
SILM
474
0
0
25 Nov 2025
Taxonomy, Evaluation and Exploitation of IPI-Centric LLM Agent Defense Frameworks
Zimo Ji
Xunguang Wang
Zongjie Li
Pingchuan Ma
Yudong Gao
Daoyuan Wu
Xincheng Yan
Tian Tian
Shuai Wang
LLMAG
AAML
289
0
0
19 Nov 2025
DRIP: Defending Prompt Injection via Token-wise Representation Editing and Residual Instruction Fusion
Ruofan Liu
Yun Lin
Zhiyong Huang
Jin Song Dong
AAML
SILM
314
0
0
01 Nov 2025
Defending Against Prompt Injection with DataFilter
Yizhu Wang
Sizhe Chen
Raghad Alkhudair
Basel Alomair
David Wagner
AAML
172
2
0
22 Oct 2025
Breaking and Fixing Defenses Against Control-Flow Hijacking in Multi-Agent Systems
Rishi Jha
Harold Triedman
Justin Wagle
Vitaly Shmatikov
AAML
117
1
0
20 Oct 2025
PIShield: Detecting Prompt Injection Attacks via Intrinsic LLM Features
Wei Zou
Yupei Liu
Yanting Wang
Ying Chen
Neil Zhenqiang Gong
Jinyuan Jia
AAML
174
0
0
15 Oct 2025
The Attacker Moves Second: Stronger Adaptive Attacks Bypass Defenses Against Llm Jailbreaks and Prompt Injections
Milad Nasr
Nicholas Carlini
Chawin Sitawarin
Sander Schulhoff
Jamie Hayes
...
Ilia Shumailov
Abhradeep Thakurta
Kai Yuanqing Xiao
Seth Neel
F. Tramèr
AAML
ELM
147
11
0
10 Oct 2025
CommandSans: Securing AI Agents with Surgical Precision Prompt Sanitization
Debeshee Das
Luca Beurer-Kellner
Marc Fischer
Maximilian Baader
AAML
125
0
0
09 Oct 2025
RL Is a Hammer and LLMs Are Nails: A Simple Reinforcement Learning Recipe for Strong Prompt Injection
Yuxin Wen
Arman Zharmagambetov
Ivan Evtimov
Narine Kokhlikyan
Tom Goldstein
Kamalika Chaudhuri
Chuan Guo
OffRL
SILM
135
4
0
06 Oct 2025
Better Privilege Separation for Agents by Restricting Data Types
Dennis Jacob
Emad Alghamdi
Zhanhao Hu
Basel Alomair
David Wagner
AAML
60
0
0
30 Sep 2025
SecInfer: Preventing Prompt Injection via Inference-time Scaling
Yupei Liu
Yanting Wang
Yuqi Jia
Jinyuan Jia
Neil Zhenqiang Gong
LRM
SILM
AAML
365
3
0
29 Sep 2025
Automatic Red Teaming LLM-based Agents with Model Context Protocol Tools
Ping He
Changjiang Li
Xingshuang Lin
Xuhong Zhang
R. Beyah
LLMAG
AAML
142
1
0
25 Sep 2025
The Sum Leaks More Than Its Parts: Compositional Privacy Risks and Mitigations in Multi-Agent Collaboration
Vaidehi Patil
Elias Stengel-Eskin
Mohit Bansal
129
1
0
16 Sep 2025
Evaluating the Robustness of Retrieval-Augmented Generation to Adversarial Evidence in the Health Domain
Shakiba Amirshahi
Amin Bigdeli
Charles L. A. Clarke
Amira Ghenai
AAML
96
1
0
04 Sep 2025
Defending Against Prompt Injection With a Few DefensiveTokens
Sizhe Chen
Yizhu Wang
Nicholas Carlini
Chawin Sitawarin
David Wagner
LLMAG
AAML
SILM
177
12
0
10 Jul 2025
RedTeamCUA: Realistic Adversarial Testing of Computer-Use Agents in Hybrid Web-OS Environments
Zeyi Liao
Jaylen Jones
Linxi Jiang
Eric Fosler-Lussier
Eric Fosler-Lussier
Yu-Chuan Su
Zhiqiang Lin
Huan Sun
ELM
371
10
0
28 May 2025
Progent: Programmable Privilege Control for LLM Agents
Tianneng Shi
Jingxuan He
Yu Yang
Hongwei Li
Linyu Wu
Wenbo Guo
Dawn Song
LLMAG
270
24
0
16 Apr 2025
1