Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2406.04031
Cited By
Jailbreak Vision Language Models via Bi-Modal Adversarial Prompt
6 June 2024
Zonghao Ying
Aishan Liu
Tianyuan Zhang
Zhengmin Yu
Siyuan Liang
Xianglong Liu
Dacheng Tao
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Jailbreak Vision Language Models via Bi-Modal Adversarial Prompt"
16 / 16 papers shown
Title
T2VShield: Model-Agnostic Jailbreak Defense for Text-to-Video Models
Siyuan Liang
Jiayang Liu
Jiecheng Zhai
Tianmeng Fang
Rongcheng Tu
A. Liu
Xiaochun Cao
Dacheng Tao
VGen
49
0
0
22 Apr 2025
BlueSuffix: Reinforced Blue Teaming for Vision-Language Models Against Jailbreak Attacks
Yunhan Zhao
Xiang Zheng
Lin Luo
Yige Li
Xingjun Ma
Yu-Gang Jiang
VLM
AAML
41
3
0
28 Oct 2024
An Image Is Worth 1000 Lies: Adversarial Transferability across Prompts on Vision-Language Models
Haochen Luo
Jindong Gu
Fengyuan Liu
Philip H. S. Torr
VLM
VPVLM
AAML
36
19
0
14 Mar 2024
Semantic Mirror Jailbreak: Genetic Algorithm Based Jailbreak Prompts Against Open-source LLMs
Xiaoxia Li
Siyuan Liang
Jiyi Zhang
Hansheng Fang
Aishan Liu
Ee-Chien Chang
51
23
0
21 Feb 2024
Visual Question Answering Instruction: Unlocking Multimodal Large Language Model To Domain-Specific Visual Multitasks
Jusung Lee
Sungguk Cha
Younghyun Lee
Cheoljong Yang
MLLM
LRM
18
2
0
13 Feb 2024
WARM: On the Benefits of Weight Averaged Reward Models
Alexandre Ramé
Nino Vieillard
Léonard Hussenot
Robert Dadashi
Geoffrey Cideron
Olivier Bachem
Johan Ferret
92
92
0
22 Jan 2024
Does Few-shot Learning Suffer from Backdoor Attacks?
Xinwei Liu
Xiaojun Jia
Jindong Gu
Yuan Xun
Siyuan Liang
Xiaochun Cao
58
18
0
31 Dec 2023
Pre-trained Trojan Attacks for Visual Recognition
Aishan Liu
Xinwei Zhang
Yisong Xiao
Yuguang Zhou
Siyuan Liang
Jiakai Wang
Xianglong Liu
Xiaochun Cao
Dacheng Tao
AAML
61
25
0
23 Dec 2023
FigStep: Jailbreaking Large Vision-Language Models via Typographic Visual Prompts
Yichen Gong
Delong Ran
Jinyuan Liu
Conglei Wang
Tianshuo Cong
Anyu Wang
Sisi Duan
Xiaoyun Wang
MLLM
127
116
0
09 Nov 2023
GPTFUZZER: Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts
Jiahao Yu
Xingwei Lin
Zheng Yu
Xinyu Xing
SILM
110
292
0
19 Sep 2023
On the Adversarial Robustness of Multi-Modal Foundation Models
Christian Schlarmann
Matthias Hein
AAML
90
84
0
21 Aug 2023
Exploring Inconsistent Knowledge Distillation for Object Detection with Data Augmentation
Jiawei Liang
Siyuan Liang
Aishan Liu
Ke Ma
Jingzhi Li
Xiaochun Cao
VLM
41
15
0
20 Sep 2022
Large Language Models are Zero-Shot Reasoners
Takeshi Kojima
S. Gu
Machel Reid
Yutaka Matsuo
Yusuke Iwasawa
ReLM
LRM
291
2,712
0
24 May 2022
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
Jason W. Wei
Xuezhi Wang
Dale Schuurmans
Maarten Bosma
Brian Ichter
F. Xia
Ed H. Chi
Quoc Le
Denny Zhou
LM&Ro
LRM
AI4CE
ReLM
315
8,261
0
28 Jan 2022
Dual Attention Suppression Attack: Generate Adversarial Camouflage in Physical World
Jiakai Wang
Aishan Liu
Zixin Yin
Shunchang Liu
Shiyu Tang
Xianglong Liu
AAML
133
191
0
01 Mar 2021
Transferable Adversarial Attacks for Image and Video Object Detection
Xingxing Wei
Siyuan Liang
Ning Chen
Xiaochun Cao
AAML
63
186
0
30 Nov 2018
1