Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1907.11932
Cited By
v1
v2
v3
v4
v5
v6 (latest)
Is BERT Really Robust? A Strong Baseline for Natural Language Attack on Text Classification and Entailment
27 July 2019
Di Jin
Zhijing Jin
Qiufeng Wang
Peter Szolovits
SILM
AAML
Re-assign community
ArXiv (abs)
PDF
HTML
Github (511★)
Papers citing
"Is BERT Really Robust? A Strong Baseline for Natural Language Attack on Text Classification and Entailment"
50 / 567 papers shown
Title
Are Language Models Agnostic to Linguistically Grounded Perturbations? A Case Study of Indic Languages
Poulami Ghosh
Raj Dabre
Pushpak Bhattacharyya
AAML
120
0
0
14 Dec 2024
BinarySelect to Improve Accessibility of Black-Box Attack Research
Shatarupa Ghosh
Jonathan Rusert
AAML
143
0
0
13 Dec 2024
Multi-Granularity Tibetan Textual Adversarial Attack Method Based on Masked Language Model
Xi Cao
Nuo Qun
Quzong Gesang
Yulei Zhu
Trashi Nyima
AAML
112
3
0
03 Dec 2024
Pay Attention to the Robustness of Chinese Minority Language Models! Syllable-level Textual Adversarial Attack on Tibetan Script
Xi Cao
Dolma Dawa
Nuo Qun
Trashi Nyima
AAML
172
4
0
03 Dec 2024
Adversarial Prompt Distillation for Vision-Language Models
Lin Luo
Xin Wang
Bojia Zi
Shihao Zhao
Xingjun Ma
Yu-Gang Jiang
AAML
VLM
171
4
0
22 Nov 2024
Who Can Withstand Chat-Audio Attacks? An Evaluation Benchmark for Large Audio-Language Models
Wanqi Yang
Yongqian Li
Meng Fang
Yunchao Wei
Dinesh Manocha
AAML
ELM
AuLLM
119
0
0
22 Nov 2024
Rethinking the Intermediate Features in Adversarial Attacks: Misleading Robotic Models via Adversarial Distillation
Ke Zhao
Huayang Huang
Miao Li
Yu Wu
AAML
110
1
0
21 Nov 2024
NMT-Obfuscator Attack: Ignore a sentence in translation with only one word
Sahar Sadrizadeh
César Descalzo
Ljiljana Dolamic
P. Frossard
AAML
115
0
0
19 Nov 2024
IAE: Irony-based Adversarial Examples for Sentiment Analysis Systems
Xiaoyin Yi
Jiacheng Huang
AAML
123
1
0
12 Nov 2024
A Comparative Analysis of Counterfactual Explanation Methods for Text Classifiers
Stephen McAleese
Mark Keane
63
0
0
04 Nov 2024
Achieving Domain-Independent Certified Robustness via Knowledge Continuity
Alan Sun
Chiyu Ma
Kenneth Ge
Soroush Vosoughi
61
1
0
03 Nov 2024
DetectRL: Benchmarking LLM-Generated Text Detection in Real-World Scenarios
Junchao Wu
Runzhe Zhan
Derek F. Wong
Shu Yang
Xinyi Yang
Yulin Yuan
Lidia S. Chao
DeLMO
192
2
0
31 Oct 2024
ProTransformer: Robustify Transformers via Plug-and-Play Paradigm
Zhichao Hou
Weizhi Gao
Yuchen Shen
Feiyi Wang
Xiaorui Liu
VLM
67
2
0
30 Oct 2024
Attacks against Abstractive Text Summarization Models through Lead Bias and Influence Functions
Poojitha Thota
Shirin Nilizadeh
69
2
0
26 Oct 2024
Advancing NLP Security by Leveraging LLMs as Adversarial Engines
Sudarshan Srinivasan
Maria Mahbub
Amir Sadovnik
AAML
44
0
0
23 Oct 2024
Unitary Multi-Margin BERT for Robust Natural Language Processing
Hao-Yuan Chang
Kang L. Wang
AAML
49
0
0
16 Oct 2024
RAFT: Realistic Attacks to Fool Text Detectors
James Wang
Ran Li
Junfeng Yang
Chengzhi Mao
AAML
DeLMO
52
4
0
04 Oct 2024
SteerDiff: Steering towards Safe Text-to-Image Diffusion Models
Hongxiang Zhang
Yifeng He
Hao Chen
82
5
0
03 Oct 2024
IndicSentEval: How Effectively do Multilingual Transformer Models encode Linguistic Properties for Indic Languages?
Akhilesh Aravapalli
Mounika Marreddy
Subba Reddy Oota
R. Mamidi
Manish Gupta
89
0
0
03 Oct 2024
Trustworthy Text-to-Image Diffusion Models: A Timely and Focused Survey
Yi Zhang
Zhen Chen
Chih-Hong Cheng
Wenjie Ruan
Xiaowei Huang
Dezong Zhao
David Flynn
Siddartha Khastgir
Xingyu Zhao
MedIm
97
4
0
26 Sep 2024
Faithfulness and the Notion of Adversarial Sensitivity in NLP Explanations
Supriya Manna
Niladri Sett
AAML
49
1
0
26 Sep 2024
ID-Free Not Risk-Free: LLM-Powered Agents Unveil Risks in ID-Free Recommender Systems
Zehua Wang
Min Gao
Junliang Yu
Xinyi Gao
Quoc Viet Hung Nguyen
S. Sadiq
Hongzhi Yin
AAML
109
3
0
18 Sep 2024
Contextual Breach: Assessing the Robustness of Transformer-based QA Models
Asir Saadat
Nahian Ibn Asad
Md Farhan Ishmam
AAML
85
0
0
17 Sep 2024
Enhancing adversarial robustness in Natural Language Inference using explanations
Alexandros Koulakos
Maria Lymperaiou
Giorgos Filandrianos
Giorgos Stamou
SILM
AAML
132
2
0
11 Sep 2024
Adversarial Attacks on Data Attribution
Xinhe Wang
Pingbang Hu
Junwei Deng
Jiaqi W. Ma
TDI
140
0
0
09 Sep 2024
OpenFact at CheckThat! 2024: Combining Multiple Attack Methods for Effective Adversarial Text Generation
Włodzimierz Lewoniewski
P. Stolarski
Milena Stróżyna
Elzbieta Lewanska
Aleksandra Wojewoda
Ewelina Księżniak
Marcin Sawiński
AAML
68
3
0
04 Sep 2024
TF-Attack: Transferable and Fast Adversarial Attacks on Large Language Models
Zelin Li
Kehai Chen
Lemao Liu
Xuefeng Bai
Mingming Yang
Yang Xiang
Min Zhang
AAML
72
1
0
26 Aug 2024
Probing the Robustness of Vision-Language Pretrained Models: A Multimodal Adversarial Attack Approach
Jiwei Guan
Tianyu Ding
Longbing Cao
Lei Pan
Chen Wang
Xi Zheng
AAML
116
2
0
24 Aug 2024
Optimal and efficient text counterfactuals using Graph Neural Networks
Dimitris Lymperopoulos
Maria Lymperaiou
Giorgos Filandrianos
Giorgos Stamou
68
1
0
04 Aug 2024
Risks, Causes, and Mitigations of Widespread Deployments of Large Language Models (LLMs): A Survey
Md. Nazmus Sakib
Md Athikul Islam
Royal Pathak
Md Mashrur Arifin
ALM
PILM
61
3
0
01 Aug 2024
CERT-ED: Certifiably Robust Text Classification for Edit Distance
Zhuoqun Huang
Yipeng Wang
Seunghee Shin
Benjamin I. P. Rubinstein
AAML
95
1
0
01 Aug 2024
Autonomous LLM-Enhanced Adversarial Attack for Text-to-Motion
Honglei Miao
Fan Ma
Ruijie Quan
Kun Zhan
Yi Yang
AAML
105
3
0
01 Aug 2024
Fuzz-Testing Meets LLM-Based Agents: An Automated and Efficient Framework for Jailbreaking Text-To-Image Generation Models
Yingkai Dong
Xiangtao Meng
Ning Yu
Zheng Li
Shanqing Guo
LLMAG
110
17
0
01 Aug 2024
Enhancing Adversarial Text Attacks on BERT Models with Projected Gradient Descent
Hetvi Waghela
Jaydip Sen
Sneha Rakshit
AAML
SILM
74
3
0
29 Jul 2024
EaTVul: ChatGPT-based Evasion Attack Against Software Vulnerability Detection
Shigang Liu
Di Cao
Junae Kim
Tamas Abraham
Paul Montague
S. Çamtepe
Jun Zhang
Yang Xiang
AAML
106
2
0
27 Jul 2024
Benchmark Granularity and Model Robustness for Image-Text Retrieval
Mariya Hendriksen
Shuo Zhang
R. Reinanda
Mohamed Yahya
Edgar Meij
Maarten de Rijke
75
0
0
21 Jul 2024
Human-Interpretable Adversarial Prompt Attack on Large Language Models with Situational Context
Nilanjana Das
Edward Raff
Manas Gaur
AAML
93
2
0
19 Jul 2024
PG-Attack: A Precision-Guided Adversarial Attack Framework Against Vision Foundation Models for Autonomous Driving
Jiyuan Fu
Zhaoyu Chen
Kaixun Jiang
Haijing Guo
Shuyong Gao
Wenqiang Zhang
AAML
81
1
0
18 Jul 2024
Rethinking Targeted Adversarial Attacks For Neural Machine Translation
Junjie Wu
Lemao Liu
Wei Bi
Dit-Yan Yeung
AAML
84
0
0
07 Jul 2024
Spiking Convolutional Neural Networks for Text Classification
Changze Lv
Jianhan Xu
Xiaoqing Zheng
93
33
0
27 Jun 2024
Unmasking Database Vulnerabilities: Zero-Knowledge Schema Inference Attacks in Text-to-SQL Systems
Đorđe Klisura
Anthony Rios
AAML
94
2
0
20 Jun 2024
MaskPure: Improving Defense Against Text Adversaries with Stochastic Purification
Harrison Gietz
Jugal Kalita
AAML
58
1
0
18 Jun 2024
Obfuscating IoT Device Scanning Activity via Adversarial Example Generation
Haocong Li
Yaxin Zhang
Long Cheng
Wenjia Niu
Haining Wang
Qiang Li
AAML
69
0
0
17 Jun 2024
MoE-RBench
\texttt{MoE-RBench}
MoE-RBench
: Towards Building Reliable Language Models with Sparse Mixture-of-Experts
Guanjie Chen
Xinyu Zhao
Tianlong Chen
Yu Cheng
MoE
116
5
0
17 Jun 2024
It Takes Two: On the Seamlessness between Reward and Policy Model in RLHF
Taiming Lu
Lingfeng Shen
Xinyu Yang
Weiting Tan
Beidi Chen
Huaxiu Yao
101
3
0
12 Jun 2024
One Perturbation is Enough: On Generating Universal Adversarial Perturbations against Vision-Language Pre-training Models
Hao Fang
Jiawei Kong
Wenbo Yu
Bin Chen
Jiawei Li
Hao Wu
Ke Xu
Ke Xu
AAML
VLM
131
13
0
08 Jun 2024
CR-UTP: Certified Robustness against Universal Text Perturbations on Large Language Models
Qian Lou
Xin Liang
Jiaqi Xue
Yancheng Zhang
Rui Xie
Mengxin Zheng
AAML
66
0
0
04 Jun 2024
Are AI-Generated Text Detectors Robust to Adversarial Perturbations?
Guanhua Huang
Yuchen Zhang
Zhe Li
Yongjian You
Mingze Wang
Zhouwang Yang
DeLMO
111
6
0
03 Jun 2024
Tool Learning with Large Language Models: A Survey
Changle Qu
Sunhao Dai
Xiaochi Wei
Hengyi Cai
Shuaiqiang Wang
Dawei Yin
Jun Xu
Jirong Wen
LLMAG
99
107
0
28 May 2024
Evaluating the Adversarial Robustness of Retrieval-Based In-Context Learning for Large Language Models
Simon Chi Lok Yu
Jie He
Pasquale Minervini
Jeff Z. Pan
52
0
0
24 May 2024
Previous
1
2
3
4
5
...
10
11
12
Next