ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2004.01970
  4. Cited By
BAE: BERT-based Adversarial Examples for Text Classification

BAE: BERT-based Adversarial Examples for Text Classification

4 April 2020
Siddhant Garg
Goutham Ramakrishnan
    AAML
    SILM
ArXivPDFHTML

Papers citing "BAE: BERT-based Adversarial Examples for Text Classification"

50 / 81 papers shown
Title
CheatAgent: Attacking LLM-Empowered Recommender Systems via LLM Agent
CheatAgent: Attacking LLM-Empowered Recommender Systems via LLM Agent
Liang-bo Ning
Shijie Wang
Wenqi Fan
Qing Li
Xin Xu
Hao Chen
Feiran Huang
AAML
26
16
0
13 Apr 2025
FitCF: A Framework for Automatic Feature Importance-guided Counterfactual Example Generation
FitCF: A Framework for Automatic Feature Importance-guided Counterfactual Example Generation
Qianli Wang
Nils Feldhus
Simon Ostermann
Luis Felipe Villa-Arenas
Sebastian Möller
Vera Schmitt
AAML
34
0
0
01 Jan 2025
Human-Readable Adversarial Prompts: An Investigation into LLM Vulnerabilities Using Situational Context
Human-Readable Adversarial Prompts: An Investigation into LLM Vulnerabilities Using Situational Context
Nilanjana Das
Edward Raff
Manas Gaur
AAML
103
1
0
20 Dec 2024
TaeBench: Improving Quality of Toxic Adversarial Examples
TaeBench: Improving Quality of Toxic Adversarial Examples
Xuan Zhu
Dmitriy Bespalov
Liwen You
Ninad Kulkarni
Yanjun Qi
AAML
63
0
0
08 Oct 2024
Reducing and Exploiting Data Augmentation Noise through Meta Reweighting
  Contrastive Learning for Text Classification
Reducing and Exploiting Data Augmentation Noise through Meta Reweighting Contrastive Learning for Text Classification
Guanyi Mou
Yichuan Li
Kyumin Lee
26
3
0
26 Sep 2024
An Adversarial Perspective on Machine Unlearning for AI Safety
An Adversarial Perspective on Machine Unlearning for AI Safety
Jakub Łucki
Boyi Wei
Yangsibo Huang
Peter Henderson
F. Tramèr
Javier Rando
MU
AAML
71
31
0
26 Sep 2024
CERT-ED: Certifiably Robust Text Classification for Edit Distance
CERT-ED: Certifiably Robust Text Classification for Edit Distance
Zhuoqun Huang
Yipeng Wang
Seunghee Shin
Benjamin I. P. Rubinstein
AAML
40
1
0
01 Aug 2024
Human-Interpretable Adversarial Prompt Attack on Large Language Models
  with Situational Context
Human-Interpretable Adversarial Prompt Attack on Large Language Models with Situational Context
Nilanjana Das
Edward Raff
Manas Gaur
AAML
35
2
0
19 Jul 2024
IDT: Dual-Task Adversarial Attacks for Privacy Protection
IDT: Dual-Task Adversarial Attacks for Privacy Protection
Pedro Faustini
Shakila Mahjabin Tonni
Annabelle McIver
Qiongkai Xu
Mark Dras
SILM
AAML
44
0
0
28 Jun 2024
SoK: Leveraging Transformers for Malware Analysis
SoK: Leveraging Transformers for Malware Analysis
Pradip Kunwar
Kshitiz Aryal
Maanak Gupta
Mahmoud Abdelsalam
Elisa Bertino
90
0
0
27 May 2024
Towards Effective Paraphrasing for Information Disguise
Towards Effective Paraphrasing for Information Disguise
Anmol Agarwal
Shrey Gupta
Vamshi Bonagiri
Manas Gaur
Joseph M. Reagle
Ponnurangam Kumaraguru
27
3
0
08 Nov 2023
Toward Stronger Textual Attack Detectors
Toward Stronger Textual Attack Detectors
Pierre Colombo
Marine Picot
Nathan Noiry
Guillaume Staerman
Pablo Piantanida
38
5
0
21 Oct 2023
A Classification-Guided Approach for Adversarial Attacks against Neural
  Machine Translation
A Classification-Guided Approach for Adversarial Attacks against Neural Machine Translation
Sahar Sadrizadeh
Ljiljana Dolamic
P. Frossard
AAML
SILM
27
2
0
29 Aug 2023
Tag Prediction of Competitive Programming Problems using Deep Learning
  Techniques
Tag Prediction of Competitive Programming Problems using Deep Learning Techniques
Taha Lokat
Divya Prajapati
Shubhada Labde
11
1
0
03 Aug 2023
LimeAttack: Local Explainable Method for Textual Hard-Label Adversarial
  Attack
LimeAttack: Local Explainable Method for Textual Hard-Label Adversarial Attack
HaiXiang Zhu
Zhaoqing Yang
Weiwei Shang
Yuren Wu
AAML
FAtt
8
3
0
01 Aug 2023
Text-CRS: A Generalized Certified Robustness Framework against Textual
  Adversarial Attacks
Text-CRS: A Generalized Certified Robustness Framework against Textual Adversarial Attacks
Xinyu Zhang
Hanbin Hong
Yuan Hong
Peng Huang
Binghui Wang
Zhongjie Ba
Kui Ren
SILM
23
18
0
31 Jul 2023
The GANfather: Controllable generation of malicious activity to improve
  defence systems
The GANfather: Controllable generation of malicious activity to improve defence systems
Ricardo Pereira
Jacopo Bono
João Tiago Ascensão
David Oliveira Aparício
Pedro Ribeiro
P. Bizarro
AAML
21
2
0
25 Jul 2023
QUERT: Continual Pre-training of Language Model for Query Understanding
  in Travel Domain Search
QUERT: Continual Pre-training of Language Model for Query Understanding in Travel Domain Search
Jian Xie
Yidan Liang
Jingping Liu
Yanghua Xiao
Baohua Wu
Shenghua Ni
VLM
LRM
30
8
0
11 Jun 2023
Symmetric Replay Training: Enhancing Sample Efficiency in Deep
  Reinforcement Learning for Combinatorial Optimization
Symmetric Replay Training: Enhancing Sample Efficiency in Deep Reinforcement Learning for Combinatorial Optimization
Hyeon-Seob Kim
Minsu Kim
Sungsoo Ahn
Jinkyoo Park
OffRL
39
7
0
02 Jun 2023
From Adversarial Arms Race to Model-centric Evaluation: Motivating a
  Unified Automatic Robustness Evaluation Framework
From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation Framework
Yangyi Chen
Hongcheng Gao
Ganqu Cui
Lifan Yuan
Dehan Kong
...
Longtao Huang
H. Xue
Zhiyuan Liu
Maosong Sun
Heng Ji
AAML
ELM
25
6
0
29 May 2023
Modeling Adversarial Attack on Pre-trained Language Models as Sequential
  Decision Making
Modeling Adversarial Attack on Pre-trained Language Models as Sequential Decision Making
Xuanjie Fang
Sijie Cheng
Yang Liu
Wen Wang
AAML
28
9
0
27 May 2023
Entailment as Robust Self-Learner
Entailment as Robust Self-Learner
Jiaxin Ge
Hongyin Luo
Yoon Kim
James R. Glass
36
3
0
26 May 2023
On Robustness of Finetuned Transformer-based NLP Models
On Robustness of Finetuned Transformer-based NLP Models
Pavan Kalyan Reddy Neerudu
S. Oota
Mounika Marreddy
Venkateswara Rao Kagita
Manish Gupta
24
7
0
23 May 2023
Assessing Hidden Risks of LLMs: An Empirical Study on Robustness,
  Consistency, and Credibility
Assessing Hidden Risks of LLMs: An Empirical Study on Robustness, Consistency, and Credibility
Wen-song Ye
Mingfeng Ou
Tianyi Li
Yipeng Chen
Xuetao Ma
...
Sai Wu
Jie Fu
Gang Chen
Haobo Wang
J. Zhao
42
36
0
15 May 2023
Attack Named Entity Recognition by Entity Boundary Interference
Attack Named Entity Recognition by Entity Boundary Interference
Yifei Yang
Hongqiu Wu
Hai Zhao
AAML
22
5
0
09 May 2023
Explaining Recommendation System Using Counterfactual Textual
  Explanations
Explaining Recommendation System Using Counterfactual Textual Explanations
Niloofar Ranjbar
S. Momtazi
MohammadMehdi Homayounpour
27
4
0
14 Mar 2023
Knowledge-Based Counterfactual Queries for Visual Question Answering
Knowledge-Based Counterfactual Queries for Visual Question Answering
Theodoti Stoikou
Maria Lymperaiou
Giorgos Stamou
AAML
26
1
0
05 Mar 2023
Data Augmentation for Neural NLP
Data Augmentation for Neural NLP
Domagoj Pluscec
Jan Snajder
8
6
0
22 Feb 2023
Auditing large language models: a three-layered approach
Auditing large language models: a three-layered approach
Jakob Mokander
Jonas Schuett
Hannah Rose Kirk
Luciano Floridi
AILaw
MLAU
39
194
0
16 Feb 2023
Step by Step Loss Goes Very Far: Multi-Step Quantization for Adversarial
  Text Attacks
Step by Step Loss Goes Very Far: Multi-Step Quantization for Adversarial Text Attacks
Piotr Gaiñski
Klaudia Bałazy
9
6
0
10 Feb 2023
TextShield: Beyond Successfully Detecting Adversarial Sentences in Text
  Classification
TextShield: Beyond Successfully Detecting Adversarial Sentences in Text Classification
Lingfeng Shen
Ze Zhang
Haiyun Jiang
Ying Chen
AAML
37
5
0
03 Feb 2023
Impact of Adversarial Training on Robustness and Generalizability of
  Language Models
Impact of Adversarial Training on Robustness and Generalizability of Language Models
Enes Altinisik
Hassan Sajjad
H. Sencar
Safa Messaoud
Sanjay Chawla
AAML
11
8
0
10 Nov 2022
Preserving Semantics in Textual Adversarial Attacks
Preserving Semantics in Textual Adversarial Attacks
David Herel
Hugo Cisneros
Tomáš Mikolov
AAML
30
6
0
08 Nov 2022
Robustness of Fusion-based Multimodal Classifiers to Cross-Modal Content
  Dilutions
Robustness of Fusion-based Multimodal Classifiers to Cross-Modal Content Dilutions
Gaurav Verma
Vishwa Vinay
Ryan A. Rossi
Srijan Kumar
VLM
AAML
11
8
0
04 Nov 2022
Character-level White-Box Adversarial Attacks against Transformers via
  Attachable Subwords Substitution
Character-level White-Box Adversarial Attacks against Transformers via Attachable Subwords Substitution
Aiwei Liu
Honghai Yu
Xuming Hu
Shuang Li
Li Lin
Fukun Ma
Yawen Yang
Lijie Wen
26
33
0
31 Oct 2022
RoChBert: Towards Robust BERT Fine-tuning for Chinese
RoChBert: Towards Robust BERT Fine-tuning for Chinese
Zihan Zhang
Jinfeng Li
Ning Shi
Bo Yuan
Xiangyu Liu
Rong Zhang
Hui Xue
Donghong Sun
Chao Zhang
AAML
24
4
0
28 Oct 2022
Secure and Trustworthy Artificial Intelligence-Extended Reality (AI-XR)
  for Metaverses
Secure and Trustworthy Artificial Intelligence-Extended Reality (AI-XR) for Metaverses
Adnan Qayyum
M. A. Butt
Hassan Ali
Muhammad Usman
O. Halabi
Ala I. Al-Fuqaha
Q. Abbasi
Muhammad Ali Imran
Junaid Qadir
28
32
0
24 Oct 2022
ADDMU: Detection of Far-Boundary Adversarial Examples with Data and
  Model Uncertainty Estimation
ADDMU: Detection of Far-Boundary Adversarial Examples with Data and Model Uncertainty Estimation
Fan Yin
Yao Li
Cho-Jui Hsieh
Kai-Wei Chang
AAML
60
4
0
22 Oct 2022
TCAB: A Large-Scale Text Classification Attack Benchmark
TCAB: A Large-Scale Text Classification Attack Benchmark
Kalyani Asthana
Zhouhang Xie
Wencong You
Adam Noack
Jonathan Brophy
Sameer Singh
Daniel Lowd
24
3
0
21 Oct 2022
Identifying Human Strategies for Generating Word-Level Adversarial
  Examples
Identifying Human Strategies for Generating Word-Level Adversarial Examples
Maximilian Mozes
Bennett Kleinberg
Lewis D. Griffin
AAML
23
1
0
20 Oct 2022
Controllable Fake Document Infilling for Cyber Deception
Controllable Fake Document Infilling for Cyber Deception
Yibo Hu
Yu Lin
Eric Parolin
Latif Khan
Kevin W. Hamlen
22
8
0
18 Oct 2022
DFX: A Low-latency Multi-FPGA Appliance for Accelerating
  Transformer-based Text Generation
DFX: A Low-latency Multi-FPGA Appliance for Accelerating Transformer-based Text Generation
Seongmin Hong
Seungjae Moon
Junsoo Kim
Sungjae Lee
Minsub Kim
Dongsoo Lee
Joo-Young Kim
64
76
0
22 Sep 2022
A Context-Aware Approach for Textual Adversarial Attack through
  Probability Difference Guided Beam Search
A Context-Aware Approach for Textual Adversarial Attack through Probability Difference Guided Beam Search
Huijun Liu
Jie Yu
Shasha Li
Jun Ma
Bin Ji
AAML
19
1
0
17 Aug 2022
QAGAN: Adversarial Approach To Learning Domain Invariant Language
  Features
QAGAN: Adversarial Approach To Learning Domain Invariant Language Features
Shubham Shrivastava
Kaiyue Wang
OOD
16
2
0
24 Jun 2022
CodeAttack: Code-Based Adversarial Attacks for Pre-trained Programming
  Language Models
CodeAttack: Code-Based Adversarial Attacks for Pre-trained Programming Language Models
Akshita Jha
Chandan K. Reddy
SILM
ELM
AAML
25
58
0
31 May 2022
Learning to Ignore Adversarial Attacks
Learning to Ignore Adversarial Attacks
Yiming Zhang
Yan Zhou
Samuel Carton
Chenhao Tan
46
2
0
23 May 2022
AEON: A Method for Automatic Evaluation of NLP Test Cases
AEON: A Method for Automatic Evaluation of NLP Test Cases
Jen-tse Huang
Jianping Zhang
Wenxuan Wang
Pinjia He
Yuxin Su
Michael R. Lyu
30
23
0
13 May 2022
A Simple Yet Efficient Method for Adversarial Word-Substitute Attack
A Simple Yet Efficient Method for Adversarial Word-Substitute Attack
Tianle Li
Yi Yang
AAML
13
0
0
07 May 2022
Don't sweat the small stuff, classify the rest: Sample Shielding to
  protect text classifiers against adversarial attacks
Don't sweat the small stuff, classify the rest: Sample Shielding to protect text classifiers against adversarial attacks
Jonathan Rusert
P. Srinivasan
AAML
19
3
0
03 May 2022
"That Is a Suspicious Reaction!": Interpreting Logits Variation to
  Detect NLP Adversarial Attacks
"That Is a Suspicious Reaction!": Interpreting Logits Variation to Detect NLP Adversarial Attacks
Edoardo Mosca
Shreyash Agarwal
Javier Rando
Georg Groh
AAML
25
30
0
10 Apr 2022
12
Next