ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1907.11932
  4. Cited By
Is BERT Really Robust? A Strong Baseline for Natural Language Attack on
  Text Classification and Entailment
v1v2v3v4v5v6 (latest)

Is BERT Really Robust? A Strong Baseline for Natural Language Attack on Text Classification and Entailment

27 July 2019
Di Jin
Zhijing Jin
Qiufeng Wang
Peter Szolovits
    SILMAAML
ArXiv (abs)PDFHTMLGithub (511★)

Papers citing "Is BERT Really Robust? A Strong Baseline for Natural Language Attack on Text Classification and Entailment"

50 / 567 papers shown
Title
Are Language Models Agnostic to Linguistically Grounded Perturbations? A
  Case Study of Indic Languages
Are Language Models Agnostic to Linguistically Grounded Perturbations? A Case Study of Indic Languages
Poulami Ghosh
Raj Dabre
Pushpak Bhattacharyya
AAML
120
0
0
14 Dec 2024
BinarySelect to Improve Accessibility of Black-Box Attack Research
BinarySelect to Improve Accessibility of Black-Box Attack Research
Shatarupa Ghosh
Jonathan Rusert
AAML
143
0
0
13 Dec 2024
Multi-Granularity Tibetan Textual Adversarial Attack Method Based on
  Masked Language Model
Multi-Granularity Tibetan Textual Adversarial Attack Method Based on Masked Language Model
Xi Cao
Nuo Qun
Quzong Gesang
Yulei Zhu
Trashi Nyima
AAML
112
3
0
03 Dec 2024
Pay Attention to the Robustness of Chinese Minority Language Models!
  Syllable-level Textual Adversarial Attack on Tibetan Script
Pay Attention to the Robustness of Chinese Minority Language Models! Syllable-level Textual Adversarial Attack on Tibetan Script
Xi Cao
Dolma Dawa
Nuo Qun
Trashi Nyima
AAML
172
4
0
03 Dec 2024
Adversarial Prompt Distillation for Vision-Language Models
Adversarial Prompt Distillation for Vision-Language Models
Lin Luo
Xin Wang
Bojia Zi
Shihao Zhao
Xingjun Ma
Yu-Gang Jiang
AAMLVLM
171
4
0
22 Nov 2024
Who Can Withstand Chat-Audio Attacks? An Evaluation Benchmark for Large Audio-Language Models
Who Can Withstand Chat-Audio Attacks? An Evaluation Benchmark for Large Audio-Language Models
Wanqi Yang
Yongqian Li
Meng Fang
Yunchao Wei
Dinesh Manocha
AAMLELMAuLLM
119
0
0
22 Nov 2024
Rethinking the Intermediate Features in Adversarial Attacks: Misleading
  Robotic Models via Adversarial Distillation
Rethinking the Intermediate Features in Adversarial Attacks: Misleading Robotic Models via Adversarial Distillation
Ke Zhao
Huayang Huang
Miao Li
Yu Wu
AAML
110
1
0
21 Nov 2024
NMT-Obfuscator Attack: Ignore a sentence in translation with only one
  word
NMT-Obfuscator Attack: Ignore a sentence in translation with only one word
Sahar Sadrizadeh
César Descalzo
Ljiljana Dolamic
P. Frossard
AAML
115
0
0
19 Nov 2024
IAE: Irony-based Adversarial Examples for Sentiment Analysis Systems
IAE: Irony-based Adversarial Examples for Sentiment Analysis Systems
Xiaoyin Yi
Jiacheng Huang
AAML
123
1
0
12 Nov 2024
A Comparative Analysis of Counterfactual Explanation Methods for Text
  Classifiers
A Comparative Analysis of Counterfactual Explanation Methods for Text Classifiers
Stephen McAleese
Mark Keane
63
0
0
04 Nov 2024
Achieving Domain-Independent Certified Robustness via Knowledge
  Continuity
Achieving Domain-Independent Certified Robustness via Knowledge Continuity
Alan Sun
Chiyu Ma
Kenneth Ge
Soroush Vosoughi
61
1
0
03 Nov 2024
DetectRL: Benchmarking LLM-Generated Text Detection in Real-World Scenarios
DetectRL: Benchmarking LLM-Generated Text Detection in Real-World Scenarios
Junchao Wu
Runzhe Zhan
Derek F. Wong
Shu Yang
Xinyi Yang
Yulin Yuan
Lidia S. Chao
DeLMO
192
2
0
31 Oct 2024
ProTransformer: Robustify Transformers via Plug-and-Play Paradigm
ProTransformer: Robustify Transformers via Plug-and-Play Paradigm
Zhichao Hou
Weizhi Gao
Yuchen Shen
Feiyi Wang
Xiaorui Liu
VLM
67
2
0
30 Oct 2024
Attacks against Abstractive Text Summarization Models through Lead Bias
  and Influence Functions
Attacks against Abstractive Text Summarization Models through Lead Bias and Influence Functions
Poojitha Thota
Shirin Nilizadeh
69
2
0
26 Oct 2024
Advancing NLP Security by Leveraging LLMs as Adversarial Engines
Advancing NLP Security by Leveraging LLMs as Adversarial Engines
Sudarshan Srinivasan
Maria Mahbub
Amir Sadovnik
AAML
44
0
0
23 Oct 2024
Unitary Multi-Margin BERT for Robust Natural Language Processing
Unitary Multi-Margin BERT for Robust Natural Language Processing
Hao-Yuan Chang
Kang L. Wang
AAML
49
0
0
16 Oct 2024
RAFT: Realistic Attacks to Fool Text Detectors
RAFT: Realistic Attacks to Fool Text Detectors
James Wang
Ran Li
Junfeng Yang
Chengzhi Mao
AAMLDeLMO
52
4
0
04 Oct 2024
SteerDiff: Steering towards Safe Text-to-Image Diffusion Models
SteerDiff: Steering towards Safe Text-to-Image Diffusion Models
Hongxiang Zhang
Yifeng He
Hao Chen
82
5
0
03 Oct 2024
IndicSentEval: How Effectively do Multilingual Transformer Models encode
  Linguistic Properties for Indic Languages?
IndicSentEval: How Effectively do Multilingual Transformer Models encode Linguistic Properties for Indic Languages?
Akhilesh Aravapalli
Mounika Marreddy
Subba Reddy Oota
R. Mamidi
Manish Gupta
89
0
0
03 Oct 2024
Trustworthy Text-to-Image Diffusion Models: A Timely and Focused Survey
Trustworthy Text-to-Image Diffusion Models: A Timely and Focused Survey
Yi Zhang
Zhen Chen
Chih-Hong Cheng
Wenjie Ruan
Xiaowei Huang
Dezong Zhao
David Flynn
Siddartha Khastgir
Xingyu Zhao
MedIm
97
4
0
26 Sep 2024
Faithfulness and the Notion of Adversarial Sensitivity in NLP
  Explanations
Faithfulness and the Notion of Adversarial Sensitivity in NLP Explanations
Supriya Manna
Niladri Sett
AAML
49
1
0
26 Sep 2024
ID-Free Not Risk-Free: LLM-Powered Agents Unveil Risks in ID-Free Recommender Systems
ID-Free Not Risk-Free: LLM-Powered Agents Unveil Risks in ID-Free Recommender Systems
Zehua Wang
Min Gao
Junliang Yu
Xinyi Gao
Quoc Viet Hung Nguyen
S. Sadiq
Hongzhi Yin
AAML
109
3
0
18 Sep 2024
Contextual Breach: Assessing the Robustness of Transformer-based QA
  Models
Contextual Breach: Assessing the Robustness of Transformer-based QA Models
Asir Saadat
Nahian Ibn Asad
Md Farhan Ishmam
AAML
85
0
0
17 Sep 2024
Enhancing adversarial robustness in Natural Language Inference using
  explanations
Enhancing adversarial robustness in Natural Language Inference using explanations
Alexandros Koulakos
Maria Lymperaiou
Giorgos Filandrianos
Giorgos Stamou
SILMAAML
132
2
0
11 Sep 2024
Adversarial Attacks on Data Attribution
Adversarial Attacks on Data Attribution
Xinhe Wang
Pingbang Hu
Junwei Deng
Jiaqi W. Ma
TDI
140
0
0
09 Sep 2024
OpenFact at CheckThat! 2024: Combining Multiple Attack Methods for
  Effective Adversarial Text Generation
OpenFact at CheckThat! 2024: Combining Multiple Attack Methods for Effective Adversarial Text Generation
Włodzimierz Lewoniewski
P. Stolarski
Milena Stróżyna
Elzbieta Lewanska
Aleksandra Wojewoda
Ewelina Księżniak
Marcin Sawiński
AAML
68
3
0
04 Sep 2024
TF-Attack: Transferable and Fast Adversarial Attacks on Large Language
  Models
TF-Attack: Transferable and Fast Adversarial Attacks on Large Language Models
Zelin Li
Kehai Chen
Lemao Liu
Xuefeng Bai
Mingming Yang
Yang Xiang
Min Zhang
AAML
72
1
0
26 Aug 2024
Probing the Robustness of Vision-Language Pretrained Models: A
  Multimodal Adversarial Attack Approach
Probing the Robustness of Vision-Language Pretrained Models: A Multimodal Adversarial Attack Approach
Jiwei Guan
Tianyu Ding
Longbing Cao
Lei Pan
Chen Wang
Xi Zheng
AAML
116
2
0
24 Aug 2024
Optimal and efficient text counterfactuals using Graph Neural Networks
Optimal and efficient text counterfactuals using Graph Neural Networks
Dimitris Lymperopoulos
Maria Lymperaiou
Giorgos Filandrianos
Giorgos Stamou
68
1
0
04 Aug 2024
Risks, Causes, and Mitigations of Widespread Deployments of Large
  Language Models (LLMs): A Survey
Risks, Causes, and Mitigations of Widespread Deployments of Large Language Models (LLMs): A Survey
Md. Nazmus Sakib
Md Athikul Islam
Royal Pathak
Md Mashrur Arifin
ALMPILM
61
3
0
01 Aug 2024
CERT-ED: Certifiably Robust Text Classification for Edit Distance
CERT-ED: Certifiably Robust Text Classification for Edit Distance
Zhuoqun Huang
Yipeng Wang
Seunghee Shin
Benjamin I. P. Rubinstein
AAML
95
1
0
01 Aug 2024
Autonomous LLM-Enhanced Adversarial Attack for Text-to-Motion
Autonomous LLM-Enhanced Adversarial Attack for Text-to-Motion
Honglei Miao
Fan Ma
Ruijie Quan
Kun Zhan
Yi Yang
AAML
105
3
0
01 Aug 2024
Fuzz-Testing Meets LLM-Based Agents: An Automated and Efficient Framework for Jailbreaking Text-To-Image Generation Models
Fuzz-Testing Meets LLM-Based Agents: An Automated and Efficient Framework for Jailbreaking Text-To-Image Generation Models
Yingkai Dong
Xiangtao Meng
Ning Yu
Zheng Li
Shanqing Guo
LLMAG
110
17
0
01 Aug 2024
Enhancing Adversarial Text Attacks on BERT Models with Projected
  Gradient Descent
Enhancing Adversarial Text Attacks on BERT Models with Projected Gradient Descent
Hetvi Waghela
Jaydip Sen
Sneha Rakshit
AAMLSILM
74
3
0
29 Jul 2024
EaTVul: ChatGPT-based Evasion Attack Against Software Vulnerability
  Detection
EaTVul: ChatGPT-based Evasion Attack Against Software Vulnerability Detection
Shigang Liu
Di Cao
Junae Kim
Tamas Abraham
Paul Montague
S. Çamtepe
Jun Zhang
Yang Xiang
AAML
106
2
0
27 Jul 2024
Benchmark Granularity and Model Robustness for Image-Text Retrieval
Benchmark Granularity and Model Robustness for Image-Text Retrieval
Mariya Hendriksen
Shuo Zhang
R. Reinanda
Mohamed Yahya
Edgar Meij
Maarten de Rijke
75
0
0
21 Jul 2024
Human-Interpretable Adversarial Prompt Attack on Large Language Models
  with Situational Context
Human-Interpretable Adversarial Prompt Attack on Large Language Models with Situational Context
Nilanjana Das
Edward Raff
Manas Gaur
AAML
93
2
0
19 Jul 2024
PG-Attack: A Precision-Guided Adversarial Attack Framework Against
  Vision Foundation Models for Autonomous Driving
PG-Attack: A Precision-Guided Adversarial Attack Framework Against Vision Foundation Models for Autonomous Driving
Jiyuan Fu
Zhaoyu Chen
Kaixun Jiang
Haijing Guo
Shuyong Gao
Wenqiang Zhang
AAML
81
1
0
18 Jul 2024
Rethinking Targeted Adversarial Attacks For Neural Machine Translation
Rethinking Targeted Adversarial Attacks For Neural Machine Translation
Junjie Wu
Lemao Liu
Wei Bi
Dit-Yan Yeung
AAML
84
0
0
07 Jul 2024
Spiking Convolutional Neural Networks for Text Classification
Spiking Convolutional Neural Networks for Text Classification
Changze Lv
Jianhan Xu
Xiaoqing Zheng
93
33
0
27 Jun 2024
Unmasking Database Vulnerabilities: Zero-Knowledge Schema Inference Attacks in Text-to-SQL Systems
Unmasking Database Vulnerabilities: Zero-Knowledge Schema Inference Attacks in Text-to-SQL Systems
Đorđe Klisura
Anthony Rios
AAML
94
2
0
20 Jun 2024
MaskPure: Improving Defense Against Text Adversaries with Stochastic
  Purification
MaskPure: Improving Defense Against Text Adversaries with Stochastic Purification
Harrison Gietz
Jugal Kalita
AAML
58
1
0
18 Jun 2024
Obfuscating IoT Device Scanning Activity via Adversarial Example
  Generation
Obfuscating IoT Device Scanning Activity via Adversarial Example Generation
Haocong Li
Yaxin Zhang
Long Cheng
Wenjia Niu
Haining Wang
Qiang Li
AAML
69
0
0
17 Jun 2024
$\texttt{MoE-RBench}$: Towards Building Reliable Language Models with
  Sparse Mixture-of-Experts
MoE-RBench\texttt{MoE-RBench}MoE-RBench: Towards Building Reliable Language Models with Sparse Mixture-of-Experts
Guanjie Chen
Xinyu Zhao
Tianlong Chen
Yu Cheng
MoE
116
5
0
17 Jun 2024
It Takes Two: On the Seamlessness between Reward and Policy Model in
  RLHF
It Takes Two: On the Seamlessness between Reward and Policy Model in RLHF
Taiming Lu
Lingfeng Shen
Xinyu Yang
Weiting Tan
Beidi Chen
Huaxiu Yao
101
3
0
12 Jun 2024
One Perturbation is Enough: On Generating Universal Adversarial Perturbations against Vision-Language Pre-training Models
One Perturbation is Enough: On Generating Universal Adversarial Perturbations against Vision-Language Pre-training Models
Hao Fang
Jiawei Kong
Wenbo Yu
Bin Chen
Jiawei Li
Hao Wu
Ke Xu
Ke Xu
AAMLVLM
131
13
0
08 Jun 2024
CR-UTP: Certified Robustness against Universal Text Perturbations on
  Large Language Models
CR-UTP: Certified Robustness against Universal Text Perturbations on Large Language Models
Qian Lou
Xin Liang
Jiaqi Xue
Yancheng Zhang
Rui Xie
Mengxin Zheng
AAML
66
0
0
04 Jun 2024
Are AI-Generated Text Detectors Robust to Adversarial Perturbations?
Are AI-Generated Text Detectors Robust to Adversarial Perturbations?
Guanhua Huang
Yuchen Zhang
Zhe Li
Yongjian You
Mingze Wang
Zhouwang Yang
DeLMO
111
6
0
03 Jun 2024
Tool Learning with Large Language Models: A Survey
Tool Learning with Large Language Models: A Survey
Changle Qu
Sunhao Dai
Xiaochi Wei
Hengyi Cai
Shuaiqiang Wang
Dawei Yin
Jun Xu
Jirong Wen
LLMAG
99
107
0
28 May 2024
Evaluating the Adversarial Robustness of Retrieval-Based In-Context
  Learning for Large Language Models
Evaluating the Adversarial Robustness of Retrieval-Based In-Context Learning for Large Language Models
Simon Chi Lok Yu
Jie He
Pasquale Minervini
Jeff Z. Pan
52
0
0
24 May 2024
Previous
12345...101112
Next