ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2107.13541
  4. Cited By
Towards Robustness Against Natural Language Word Substitutions

Towards Robustness Against Natural Language Word Substitutions

International Conference on Learning Representations (ICLR), 2021
28 July 2021
Xinshuai Dong
Anh Tuan Luu
Rongrong Ji
Hong Liu
    SILMAAML
ArXiv (abs)PDFHTML

Papers citing "Towards Robustness Against Natural Language Word Substitutions"

50 / 70 papers shown
Adversarial Defence without Adversarial Defence: Enhancing Language Model Robustness via Instance-level Principal Component Removal
Adversarial Defence without Adversarial Defence: Enhancing Language Model Robustness via Instance-level Principal Component Removal
Yang Wang
Chenghao Xiao
Yi Zhou
Stuart E. Middleton
Noura Al Moubayed
C. D. Lin
AAML
303
1
0
29 Jul 2025
Bridging Robustness and Generalization Against Word Substitution Attacks in NLP via the Growth Bound Matrix Approach
Bridging Robustness and Generalization Against Word Substitution Attacks in NLP via the Growth Bound Matrix ApproachAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Mohammed Bouri
Adnane Saoud
AAMLSILM
217
0
0
14 Jul 2025
Rapid Urban Visibility Hotspots: Quantifying Building Vertex Visibility from Connected Vehicle Trajectories using Spatial Indexing
Rapid Urban Visibility Hotspots: Quantifying Building Vertex Visibility from Connected Vehicle Trajectories using Spatial Indexing
Artur Grigorev
Adriana-Simona Mihaita
298
0
0
03 Jun 2025
The Counting Power of Transformers
The Counting Power of Transformers
Marco Sälzer
Chris Köcher
Anthony Widjaja Lin
Georg Zetzsche
Anthony Widjaja Lin
375
0
0
16 May 2025
Model Hemorrhage and the Robustness Limits of Large Language Models
Model Hemorrhage and the Robustness Limits of Large Language Models
Ziyang Ma
Hui Yuan
Guang Dai
Gui-Song Xia
Bo Du
Liangpei Zhang
Dacheng Tao
317
1
0
31 Mar 2025
Confidence Elicitation: A New Attack Vector for Large Language Models
Confidence Elicitation: A New Attack Vector for Large Language ModelsInternational Conference on Learning Representations (ICLR), 2025
Brian Formento
Chuan-Sheng Foo
See-Kiong Ng
AAML
573
2
0
07 Feb 2025
Tougher Text, Smarter Models: Raising the Bar for Adversarial Defence Benchmarks
Tougher Text, Smarter Models: Raising the Bar for Adversarial Defence BenchmarksInternational Conference on Computational Linguistics (COLING), 2025
Yang Wang
Chenghua Lin
ELM
374
4
0
05 Jan 2025
ProTransformer: Robustify Transformers via Plug-and-Play Paradigm
ProTransformer: Robustify Transformers via Plug-and-Play ParadigmNeural Information Processing Systems (NeurIPS), 2024
Zhichao Hou
Weizhi Gao
Yuchen Shen
Feiyi Wang
Xiaorui Liu
VLM
259
2
0
30 Oct 2024
A New Benchmark Dataset and Mixture-of-Experts Language Models for Adversarial Natural Language Inference in Vietnamese
A New Benchmark Dataset and Mixture-of-Experts Language Models for Adversarial Natural Language Inference in Vietnamese
Tin Van Huynh
Kiet Van Nguyen
Ngan Luu-Thuy Nguyen
338
2
0
25 Jun 2024
Transformer Encoder Satisfiability: Complexity and Impact on Formal Reasoning
Transformer Encoder Satisfiability: Complexity and Impact on Formal Reasoning
Marco Sälzer
Eric Alsmann
Martin Lange
LRM
209
3
0
28 May 2024
GenFighter: A Generative and Evolutive Textual Attack Removal
GenFighter: A Generative and Evolutive Textual Attack Removal
Md Athikul Islam
Edoardo Serra
Sushil Jajodia
AAML
168
0
0
17 Apr 2024
SemRoDe: Macro Adversarial Training to Learn Representations That are
  Robust to Word-Level Attacks
SemRoDe: Macro Adversarial Training to Learn Representations That are Robust to Word-Level Attacks
Brian Formento
Wenjie Feng
Chuan-Sheng Foo
Anh Tuan Luu
See-Kiong Ng
AAML
324
9
0
27 Mar 2024
Extreme Miscalibration and the Illusion of Adversarial Robustness
Extreme Miscalibration and the Illusion of Adversarial Robustness
Vyas Raina
Samson Tan
Volkan Cevher
Aditya Rawal
Sheng Zha
George Karypis
AAML
328
4
0
27 Feb 2024
Defending Against Weight-Poisoning Backdoor Attacks for
  Parameter-Efficient Fine-Tuning
Defending Against Weight-Poisoning Backdoor Attacks for Parameter-Efficient Fine-Tuning
Shuai Zhao
Yaoyao Yu
Anh Tuan Luu
Jie Fu
Lingjuan Lyu
Meihuizi Jia
Jinming Wen
AAML
364
30
0
19 Feb 2024
Fast Adversarial Training against Textual Adversarial Attacks
Fast Adversarial Training against Textual Adversarial Attacks
Yichen Yang
Xin Liu
Kun He
AAML
178
6
0
23 Jan 2024
Toward Stronger Textual Attack Detectors
Toward Stronger Textual Attack DetectorsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Pierre Colombo
Marine Picot
Nathan Noiry
Guillaume Staerman
Pablo Piantanida
814
5
0
21 Oct 2023
Beyond Hard Samples: Robust and Effective Grammatical Error Correction
  with Cycle Self-Augmenting
Beyond Hard Samples: Robust and Effective Grammatical Error Correction with Cycle Self-Augmenting
Zecheng Tang
Kaiqi Feng
Juntao Li
Min Zhang
238
2
0
20 Oct 2023
Fooling the Textual Fooler via Randomizing Latent Representations
Fooling the Textual Fooler via Randomizing Latent RepresentationsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Duy C. Hoang
Quang H. Nguyen
Saurav Manchanda
MinLong Peng
Kok-Seng Wong
Khoa D. Doan
SILMAAML
263
2
0
02 Oct 2023
Defending Against Alignment-Breaking Attacks via Robustly Aligned LLM
Defending Against Alignment-Breaking Attacks via Robustly Aligned LLMAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Bochuan Cao
Yu Cao
Lu Lin
Jinghui Chen
AAML
456
198
0
18 Sep 2023
LimeAttack: Local Explainable Method for Textual Hard-Label Adversarial
  Attack
LimeAttack: Local Explainable Method for Textual Hard-Label Adversarial AttackAAAI Conference on Artificial Intelligence (AAAI), 2023
HaiXiang Zhu
Zhaoqing Yang
Weiwei Shang
Yuren Wu
AAMLFAtt
314
14
0
01 Aug 2023
Text-CRS: A Generalized Certified Robustness Framework against Textual
  Adversarial Attacks
Text-CRS: A Generalized Certified Robustness Framework against Textual Adversarial AttacksIEEE Symposium on Security and Privacy (IEEE S&P), 2023
Xinyu Zhang
Hanbin Hong
Yuan Hong
Peng Huang
Binghui Wang
Zhongjie Ba
Kui Ren
SILM
366
34
0
31 Jul 2023
Transferable Adversarial Robustness for Categorical Data via Universal
  Robust Embeddings
Transferable Adversarial Robustness for Categorical Data via Universal Robust EmbeddingsNeural Information Processing Systems (NeurIPS), 2023
Klim Kireev
Maksym Andriushchenko
Carmela Troncoso
Nicolas Flammarion
OOD
289
2
0
06 Jun 2023
A Causal View of Entity Bias in (Large) Language Models
A Causal View of Entity Bias in (Large) Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Fei Wang
Wen-An Mo
Yiwei Wang
Wenxuan Zhou
Muhao Chen
220
20
0
24 May 2023
Randomized Smoothing with Masked Inference for Adversarially Robust Text
  Classifications
Randomized Smoothing with Masked Inference for Adversarially Robust Text ClassificationsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Han Cheol Moon
Shafiq Joty
Ruochen Zhao
Megh Thakkar
Xu Chi
AAML
236
18
0
11 May 2023
Toward Adversarial Training on Contextualized Language Representation
Toward Adversarial Training on Contextualized Language RepresentationInternational Conference on Learning Representations (ICLR), 2023
Hongqiu Wu
Wenshu Fan
Han Shi
Haizhen Zhao
Hao Fei
AAML
154
15
0
08 May 2023
The Best Defense is Attack: Repairing Semantics in Textual Adversarial
  Examples
The Best Defense is Attack: Repairing Semantics in Textual Adversarial ExamplesConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Heng Yang
Ke Li
AAML
303
4
0
06 May 2023
ANTONIO: Towards a Systematic Method of Generating NLP Benchmarks for
  Verification
ANTONIO: Towards a Systematic Method of Generating NLP Benchmarks for Verification
Marco Casadio
Luca Arnaboldi
M. Daggitt
Omri Isac
Tanvi Dinkar
Daniel Kienitz
Verena Rieser
Ekaterina Komendantskaya
253
5
0
06 May 2023
Prompt as Triggers for Backdoor Attack: Examining the Vulnerability in
  Language Models
Prompt as Triggers for Backdoor Attack: Examining the Vulnerability in Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Shuai Zhao
Jinming Wen
Anh Tuan Luu
Jiaqi Zhao
Jie Fu
SILM
618
129
0
02 May 2023
RoCOCO: Robustness Benchmark of MS-COCO to Stress-test Image-Text
  Matching Models
RoCOCO: Robustness Benchmark of MS-COCO to Stress-test Image-Text Matching Models
Seulki Park
Daeho Um
Hajung Yoon
Sanghyuk Chun
Sangdoo Yun
Hawook Jeong
396
5
0
21 Apr 2023
Masked Language Model Based Textual Adversarial Example Detection
Masked Language Model Based Textual Adversarial Example DetectionACM Asia Conference on Computer and Communications Security (AsiaCCS), 2023
Xiaomei Zhang
Zhaoxi Zhang
Qi Zhong
Xufei Zheng
Yanjun Zhang
Shengshan Hu
L. Zhang
AAML
339
8
0
18 Apr 2023
Backdoor Learning for NLP: Recent Advances, Challenges, and Future
  Research Directions
Backdoor Learning for NLP: Recent Advances, Challenges, and Future Research Directions
Marwan Omar
SILMAAML
199
21
0
14 Feb 2023
Less is More: Understanding Word-level Textual Adversarial Attack via
  n-gram Frequency Descend
Less is More: Understanding Word-level Textual Adversarial Attack via n-gram Frequency Descend
Ning Lu
Shengcai Liu
Zhirui Zhang
Qi. Wang
Haifeng Liu
Jiaheng Zhang
AAML
341
13
0
06 Feb 2023
TextShield: Beyond Successfully Detecting Adversarial Sentences in Text
  Classification
TextShield: Beyond Successfully Detecting Adversarial Sentences in Text ClassificationInternational Conference on Learning Representations (ICLR), 2023
Lingfeng Shen
Ze Zhang
Haiyun Jiang
Ying-Cong Chen
AAML
338
9
0
03 Feb 2023
On Robustness of Prompt-based Semantic Parsing with Large Pre-trained
  Language Model: An Empirical Study on Codex
On Robustness of Prompt-based Semantic Parsing with Large Pre-trained Language Model: An Empirical Study on CodexConference of the European Chapter of the Association for Computational Linguistics (EACL), 2023
Terry Yue Zhuo
Zhuang Li
Yujin Huang
Fatemeh Shiri
Weiqing Wang
Gholamreza Haffari
Yuan-Fang Li
AAML
414
65
0
30 Jan 2023
TextGrad: Advancing Robustness Evaluation in NLP by Gradient-Driven
  Optimization
TextGrad: Advancing Robustness Evaluation in NLP by Gradient-Driven OptimizationInternational Conference on Learning Representations (ICLR), 2022
Bairu Hou
Jinghan Jia
Yihua Zhang
Guanhua Zhang
Yang Zhang
Sijia Liu
Shiyu Chang
SILMAAML
172
27
0
19 Dec 2022
Preserving Semantics in Textual Adversarial Attacks
Preserving Semantics in Textual Adversarial AttacksEuropean Conference on Artificial Intelligence (ECAI), 2022
David Herel
Hugo Cisneros
Tomas Mikolov
AAML
237
9
0
08 Nov 2022
Textual Manifold-based Defense Against Natural Language Adversarial
  Examples
Textual Manifold-based Defense Against Natural Language Adversarial ExamplesConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
D. M. Nguyen
Anh Tuan Luu
AAML
265
27
0
05 Nov 2022
Emergent Linguistic Structures in Neural Networks are Fragile
Emergent Linguistic Structures in Neural Networks are Fragile
Emanuele La Malfa
Matthew Wicker
Marta Kiatkowska
612
1
0
31 Oct 2022
Disentangled Text Representation Learning with Information-Theoretic
  Perspective for Adversarial Robustness
Disentangled Text Representation Learning with Information-Theoretic Perspective for Adversarial RobustnessIEEE/ACM Transactions on Audio Speech and Language Processing (TASLP), 2022
Jiahao Zhao
Wenji Mao
DRLOOD
180
7
0
26 Oct 2022
TCAB: A Large-Scale Text Classification Attack Benchmark
TCAB: A Large-Scale Text Classification Attack Benchmark
Kalyani Asthana
Zhouhang Xie
Wencong You
Adam Noack
Jonathan Brophy
Sameer Singh
Daniel Lowd
305
3
0
21 Oct 2022
Probabilistic Categorical Adversarial Attack & Adversarial Training
Probabilistic Categorical Adversarial Attack & Adversarial Training
Han Xu
Penghei He
Jie Ren
Yuxuan Wan
Zitao Liu
Hui Liu
Shucheng Zhou
AAMLSILM
207
1
0
17 Oct 2022
Rethinking Textual Adversarial Defense for Pre-trained Language Models
Rethinking Textual Adversarial Defense for Pre-trained Language ModelsIEEE/ACM Transactions on Audio Speech and Language Processing (TASLP), 2022
Jiayi Wang
Rongzhou Bao
Zhuosheng Zhang
Hai Zhao
AAMLSILM
201
14
0
21 Jul 2022
Certified Robustness Against Natural Language Attacks by Causal
  Intervention
Certified Robustness Against Natural Language Attacks by Causal InterventionInternational Conference on Machine Learning (ICML), 2022
Haiteng Zhao
Chang Ma
Xinshuai Dong
Anh Tuan Luu
Zhi-Hong Deng
Hanwang Zhang
AAML
357
40
0
24 May 2022
A Word is Worth A Thousand Dollars: Adversarial Attack on Tweets Fools
  Stock Predictions
A Word is Worth A Thousand Dollars: Adversarial Attack on Tweets Fools Stock PredictionsNorth American Chapter of the Association for Computational Linguistics (NAACL), 2022
Yong Xie
Dakuo Wang
Pin-Yu Chen
Jinjun Xiong
Sijia Liu
Oluwasanmi Koyejo
AAML
302
26
0
01 May 2022
Improving robustness of language models from a geometry-aware
  perspective
Improving robustness of language models from a geometry-aware perspectiveFindings (Findings), 2022
Bin Zhu
Zhaoquan Gu
Le Wang
Jinyin Chen
Qi Xuan
AAML
136
10
0
28 Apr 2022
"That Is a Suspicious Reaction!": Interpreting Logits Variation to
  Detect NLP Adversarial Attacks
"That Is a Suspicious Reaction!": Interpreting Logits Variation to Detect NLP Adversarial AttacksAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Edoardo Mosca
Shreyash Agarwal
Javier Rando
Georg Groh
AAML
216
39
0
10 Apr 2022
Text Adversarial Purification as Defense against Adversarial Attacks
Text Adversarial Purification as Defense against Adversarial AttacksAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Linyang Li
Demin Song
Xipeng Qiu
AAML
167
20
0
27 Mar 2022
Adversarial Training for Improving Model Robustness? Look at Both
  Prediction and Interpretation
Adversarial Training for Improving Model Robustness? Look at Both Prediction and InterpretationAAAI Conference on Artificial Intelligence (AAAI), 2022
Hanjie Chen
Yangfeng Ji
OODAAMLVLM
221
26
0
23 Mar 2022
On Robust Prefix-Tuning for Text Classification
On Robust Prefix-Tuning for Text ClassificationInternational Conference on Learning Representations (ICLR), 2022
Zonghan Yang
Yang Liu
VLM
179
22
0
19 Mar 2022
A Survey of Adversarial Defences and Robustness in NLP
A Survey of Adversarial Defences and Robustness in NLP
Shreyansh Goyal
Sumanth Doddapaneni
Mitesh M.Khapra
B. Ravindran
AAML
477
35
0
12 Mar 2022
12
Next