Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1704.08006
Cited By
Deep Text Classification Can be Fooled
26 April 2017
Bin Liang
Hongcheng Li
Miaoqiang Su
Pan Bian
Xirong Li
Wenchang Shi
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Deep Text Classification Can be Fooled"
50 / 58 papers shown
Title
Spiking Convolutional Neural Networks for Text Classification
Changze Lv
Jianhan Xu
Xiaoqing Zheng
56
27
0
27 Jun 2024
Adversarial Testing for Visual Grounding via Image-Aware Property Reduction
Zhiyuan Chang
Mingyang Li
Junjie Wang
Cheng Li
Boyu Wu
Fanjiang Xu
Qing Wang
AAML
36
0
0
02 Mar 2024
Trustworthy Distributed AI Systems: Robustness, Privacy, and Governance
Wenqi Wei
Ling Liu
31
16
0
02 Feb 2024
A Critical Reflection on the Use of Toxicity Detection Algorithms in Proactive Content Moderation Systems
Mark Warner
Angelika Strohmayer
Matthew Higgs
Lynne Coventry
16
5
0
19 Jan 2024
From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation Framework
Yangyi Chen
Hongcheng Gao
Ganqu Cui
Lifan Yuan
Dehan Kong
...
Longtao Huang
H. Xue
Zhiyuan Liu
Maosong Sun
Heng Ji
AAML
ELM
27
6
0
29 May 2023
Modeling Adversarial Attack on Pre-trained Language Models as Sequential Decision Making
Xuanjie Fang
Sijie Cheng
Yang Liu
Wen Wang
AAML
34
9
0
27 May 2023
Another Dead End for Morphological Tags? Perturbed Inputs and Parsing
Alberto Muñoz-Ortiz
David Vilares
30
1
0
24 May 2023
A Survey of Safety and Trustworthiness of Large Language Models through the Lens of Verification and Validation
Xiaowei Huang
Wenjie Ruan
Wei Huang
Gao Jin
Yizhen Dong
...
Sihao Wu
Peipei Xu
Dengyu Wu
André Freitas
Mustafa A. Mustafa
ALM
39
82
0
19 May 2023
Adversarial Amendment is the Only Force Capable of Transforming an Enemy into a Friend
Chong Yu
Tao Chen
Zhongxue Gan
AAML
15
1
0
18 May 2023
Tell Model Where to Attend: Improving Interpretability of Aspect-Based Sentiment Classification via Small Explanation Annotations
Zhenxiao Cheng
Jie Zhou
Wen Wu
Qin Chen
Liang He
29
3
0
21 Feb 2023
TASA: Deceiving Question Answering Models by Twin Answer Sentences Attack
Yu Cao
Dianqi Li
Meng Fang
Dinesh Manocha
Jun Gao
Yibing Zhan
Dacheng Tao
AAML
26
15
0
27 Oct 2022
ROSE: Robust Selective Fine-tuning for Pre-trained Language Models
Lan Jiang
Hao Zhou
Yankai Lin
Peng Li
Jie Zhou
R. Jiang
AAML
37
8
0
18 Oct 2022
Adversarial Robustness for Tabular Data through Cost and Utility Awareness
Klim Kireev
B. Kulynych
Carmela Troncoso
AAML
26
16
0
27 Aug 2022
A Context-Aware Approach for Textual Adversarial Attack through Probability Difference Guided Beam Search
Huijun Liu
Jie Yu
Shasha Li
Jun Ma
Bin Ji
AAML
38
1
0
17 Aug 2022
Rethinking Textual Adversarial Defense for Pre-trained Language Models
Jiayi Wang
Rongzhou Bao
Zhuosheng Zhang
Hai Zhao
AAML
SILM
17
11
0
21 Jul 2022
AEON: A Method for Automatic Evaluation of NLP Test Cases
Jen-tse Huang
Jianping Zhang
Wenxuan Wang
Pinjia He
Yuxin Su
Michael R. Lyu
40
23
0
13 May 2022
Testing the limits of natural language models for predicting human language judgments
Tal Golan
Matthew Siegelman
N. Kriegeskorte
Christopher A. Baldassano
22
15
0
07 Apr 2022
Adversarial Training for Improving Model Robustness? Look at Both Prediction and Interpretation
Hanjie Chen
Yangfeng Ji
OOD
AAML
VLM
24
21
0
23 Mar 2022
Defending Black-box Skeleton-based Human Activity Classifiers
He-Nan Wang
Yunfeng Diao
Zichang Tan
G. Guo
AAML
51
10
0
09 Mar 2022
How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial Robustness?
Xinhsuai Dong
Anh Tuan Luu
Min-Bin Lin
Shuicheng Yan
Hanwang Zhang
SILM
AAML
20
55
0
22 Dec 2021
The King is Naked: on the Notion of Robustness for Natural Language Processing
Emanuele La Malfa
Marta Z. Kwiatkowska
20
28
0
13 Dec 2021
Adversarial Attacks Against Deep Generative Models on Data: A Survey
Hui Sun
Tianqing Zhu
Zhiqiu Zhang
Dawei Jin
Wanlei Zhou
AAML
37
42
0
01 Dec 2021
Effective and Imperceptible Adversarial Textual Attack via Multi-objectivization
Shengcai Liu
Ning Lu
W. Hong
Chao Qian
Ke Tang
AAML
14
15
0
02 Nov 2021
Detecting Textual Adversarial Examples through Randomized Substitution and Vote
Xiaosen Wang
Yifeng Xiong
Kun He
AAML
17
11
0
13 Sep 2021
A Strong Baseline for Query Efficient Attacks in a Black Box Setting
Rishabh Maheshwary
Saket Maheshwary
Vikram Pudi
AAML
27
30
0
10 Sep 2021
Multi-granularity Textual Adversarial Attack with Behavior Cloning
Yangyi Chen
Jingtong Su
Wei Wei
AAML
17
32
0
09 Sep 2021
Evaluating the Robustness of Neural Language Models to Input Perturbations
M. Moradi
Matthias Samwald
AAML
48
95
0
27 Aug 2021
Towards Robustness Against Natural Language Word Substitutions
Xinshuai Dong
A. Luu
Rongrong Ji
Hong Liu
SILM
AAML
32
113
0
28 Jul 2021
We Can Always Catch You: Detecting Adversarial Patched Objects WITH or WITHOUT Signature
Binxiu Liang
Jiachun Li
Jianjun Huang
AAML
25
12
0
09 Jun 2021
Defending Against Backdoor Attacks in Natural Language Generation
Xiaofei Sun
Xiaoya Li
Yuxian Meng
Xiang Ao
Fei Wu
Jiwei Li
Tianwei Zhang
AAML
SILM
31
47
0
03 Jun 2021
Robustness Tests of NLP Machine Learning Models: Search and Semantically Replace
Rahul Singh
Karan Jindal
Yufei Yu
Hanyu Yang
Tarun Joshi
Matthew A. Campbell
Wayne B. Shoumaker
50
2
0
20 Apr 2021
BERT: A Review of Applications in Natural Language Processing and Understanding
M. V. Koroteev
VLM
25
196
0
22 Mar 2021
Adversarial Attack on Network Embeddings via Supervised Network Poisoning
Viresh Gupta
Tanmoy Chakraborty
AAML
28
12
0
14 Feb 2021
Generating Natural Language Attacks in a Hard Label Black Box Setting
Rishabh Maheshwary
Saket Maheshwary
Vikram Pudi
AAML
22
103
0
29 Dec 2020
A Deep Marginal-Contrastive Defense against Adversarial Attacks on 1D Models
Mohammed Hassanin
Nour Moustafa
M. Tahtali
AAML
22
2
0
08 Dec 2020
Self-Explaining Structures Improve NLP Models
Zijun Sun
Chun Fan
Qinghong Han
Xiaofei Sun
Yuxian Meng
Fei Wu
Jiwei Li
MILM
XAI
LRM
FAtt
39
38
0
03 Dec 2020
Do We Need Online NLU Tools?
Petr Lorenc
Petro Marek
Jan Pichl
Jakub Konrád
Jan Sedivý
13
6
0
19 Nov 2020
Can Adversarial Weight Perturbations Inject Neural Backdoors?
Siddhant Garg
Adarsh Kumar
Vibhor Goel
Yingyu Liang
AAML
39
86
0
04 Aug 2020
Defense against Adversarial Attacks in NLP via Dirichlet Neighborhood Ensemble
Yi Zhou
Xiaoqing Zheng
Cho-Jui Hsieh
Kai-Wei Chang
Xuanjing Huang
SILM
39
48
0
20 Jun 2020
Differentiable Language Model Adversarial Attacks on Categorical Sequence Classifiers
I. Fursov
A. Zaytsev
Nikita Klyuchnikov
A. Kravchenko
E. Burnaev
AAML
SILM
29
5
0
19 Jun 2020
Chat as Expected: Learning to Manipulate Black-box Neural Dialogue Models
Haochen Liu
Zhiwei Wang
Tyler Derr
Jiliang Tang
AAML
22
15
0
27 May 2020
Frequency-Guided Word Substitutions for Detecting Textual Adversarial Examples
Maximilian Mozes
Pontus Stenetorp
Bennett Kleinberg
Lewis D. Griffin
AAML
25
99
0
13 Apr 2020
Generating Natural Language Adversarial Examples on a Large Scale with Generative Models
Yankun Ren
J. Lin
Siliang Tang
Jun Zhou
Shuang Yang
Yuan Qi
Xiang Ren
GAN
AAML
SILM
16
21
0
10 Mar 2020
Adv-BERT: BERT is not robust on misspellings! Generating nature adversarial samples on BERT
Lichao Sun
Kazuma Hashimoto
Wenpeng Yin
Akari Asai
Jia Li
Philip Yu
Caiming Xiong
SILM
AAML
12
101
0
27 Feb 2020
Adversarial Robustness for Code
Pavol Bielik
Martin Vechev
AAML
14
89
0
11 Feb 2020
Say What I Want: Towards the Dark Side of Neural Dialogue Models
Haochen Liu
Tyler Derr
Zitao Liu
Jiliang Tang
28
16
0
13 Sep 2019
Negative Training for Neural Dialogue Response Generation
Tianxing He
James R. Glass
30
59
0
06 Mar 2019
Re-evaluating ADEM: A Deeper Look at Scoring Dialogue Responses
Ananya B. Sai
Mithun Das Gupta
Mitesh M. Khapra
Mukundhan Srinivasan
25
48
0
23 Feb 2019
Adversarial Attacks on Deep Learning Models in Natural Language Processing: A Survey
W. Zhang
Quan Z. Sheng
A. Alhazmi
Chenliang Li
AAML
24
57
0
21 Jan 2019
TextBugger: Generating Adversarial Text Against Real-world Applications
Jinfeng Li
S. Ji
Tianyu Du
Bo Li
Ting Wang
SILM
AAML
46
723
0
13 Dec 2018
1
2
Next