Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2407.03391
Cited By
Soft Begging: Modular and Efficient Shielding of LLMs against Prompt Injection and Jailbreaking based on Prompt Tuning
3 July 2024
Simon Ostermann
Kevin Baum
Christoph Endres
Julia Masloh
P. Schramowski
AAML
Re-assign community
ArXiv (abs)
PDF
HTML
Github
Papers citing
"Soft Begging: Modular and Efficient Shielding of LLMs against Prompt Injection and Jailbreaking based on Prompt Tuning"
3 / 3 papers shown
Detection Method for Prompt Injection by Integrating Pre-trained Model and Heuristic Feature Engineering
Yi Ji
Runzhi Li
Baolei Mao
AAML
212
1
0
05 Jun 2025
Can LLMs Separate Instructions From Data? And What Do We Even Mean By That?
International Conference on Learning Representations (ICLR), 2024
Egor Zverev
Sahar Abdelnabi
Soroush Tabesh
Mario Fritz
Christoph H. Lampert
483
54
0
11 Mar 2024
Hijacking Large Language Models via Adversarial In-Context Learning
Yao Qiang
Xiangyu Zhou
Saleh Zare Zade
Prashant Khanduri
Dongxiao Zhu
639
50
0
16 Nov 2023
1
Page 1 of 1