ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2106.09898
  4. Cited By
Bad Characters: Imperceptible NLP Attacks

Bad Characters: Imperceptible NLP Attacks

18 June 2021
Nicholas Boucher
Ilia Shumailov
Ross J. Anderson
Nicolas Papernot
    AAML
    SILM
ArXivPDFHTML

Papers citing "Bad Characters: Imperceptible NLP Attacks"

50 / 68 papers shown
Title
Sponge Attacks on Sensing AI: Energy-Latency Vulnerabilities and Defense via Model Pruning
Sponge Attacks on Sensing AI: Energy-Latency Vulnerabilities and Defense via Model Pruning
Syed Mhamudul Hasan
Hussein Zangoti
Iraklis Anagnostopoulos
Abdur R. Shahid
AAML
29
0
0
09 May 2025
A Comprehensive Analysis of Adversarial Attacks against Spam Filters
A Comprehensive Analysis of Adversarial Attacks against Spam Filters
Esra Hotoğlu
Sevil Sen
Burcu Can
AAML
26
0
0
04 May 2025
Exploring Gradient-Guided Masked Language Model to Detect Textual Adversarial Attacks
Exploring Gradient-Guided Masked Language Model to Detect Textual Adversarial Attacks
Xiaomei Zhang
Zhaoxi Zhang
Yanjun Zhang
Xufei Zheng
L. Zhang
Shengshan Hu
Shirui Pan
AAML
27
0
0
08 Apr 2025
A Grey-box Text Attack Framework using Explainable AI
Esther Chiramal
Kelvin Soh Boon Kai
AAML
SILM
57
0
0
11 Mar 2025
BiMarker: Enhancing Text Watermark Detection for Large Language Models with Bipolar Watermarks
BiMarker: Enhancing Text Watermark Detection for Large Language Models with Bipolar Watermarks
Zhuang Li
48
1
0
21 Jan 2025
The Best Defense is a Good Offense: Countering LLM-Powered Cyberattacks
The Best Defense is a Good Offense: Countering LLM-Powered Cyberattacks
Daniel Ayzenshteyn
Roy Weiss
Yisroel Mirsky
AAML
31
0
0
20 Oct 2024
Golyadkin's Torment: Doppelgängers and Adversarial Vulnerability
Golyadkin's Torment: Doppelgängers and Adversarial Vulnerability
George I. Kamberov
AAML
19
0
0
17 Oct 2024
Probing the Robustness of Vision-Language Pretrained Models: A
  Multimodal Adversarial Attack Approach
Probing the Robustness of Vision-Language Pretrained Models: A Multimodal Adversarial Attack Approach
Jiwei Guan
Tianyu Ding
Longbing Cao
Lei Pan
Chen Wang
Xi Zheng
AAML
33
1
0
24 Aug 2024
Breaking Agents: Compromising Autonomous LLM Agents Through Malfunction
  Amplification
Breaking Agents: Compromising Autonomous LLM Agents Through Malfunction Amplification
Boyang Zhang
Yicong Tan
Yun Shen
Ahmed Salem
Michael Backes
Savvas Zannettou
Yang Zhang
LLMAG
AAML
44
14
0
30 Jul 2024
Watermark Smoothing Attacks against Language Models
Watermark Smoothing Attacks against Language Models
Hongyan Chang
Hamed Hassani
Reza Shokri
WaLM
63
2
0
19 Jul 2024
SpamDam: Towards Privacy-Preserving and Adversary-Resistant SMS Spam
  Detection
SpamDam: Towards Privacy-Preserving and Adversary-Resistant SMS Spam Detection
Yekai Li
Rufan Zhang
Wenxin Rong
Xianghang Mi
34
2
0
15 Apr 2024
The Impact of Uniform Inputs on Activation Sparsity and Energy-Latency
  Attacks in Computer Vision
The Impact of Uniform Inputs on Activation Sparsity and Energy-Latency Attacks in Computer Vision
Andreas Müller
Erwin Quiring
AAML
27
1
0
27 Mar 2024
A Modified Word Saliency-Based Adversarial Attack on Text Classification
  Models
A Modified Word Saliency-Based Adversarial Attack on Text Classification Models
Hetvi Waghela
Sneha Rakshit
Jaydip Sen
AAML
18
7
0
17 Mar 2024
Architectural Neural Backdoors from First Principles
Architectural Neural Backdoors from First Principles
Harry Langford
Ilia Shumailov
Yiren Zhao
Robert D. Mullins
Nicolas Papernot
AAML
32
2
0
10 Feb 2024
Comprehensive Assessment of Jailbreak Attacks Against LLMs
Comprehensive Assessment of Jailbreak Attacks Against LLMs
Junjie Chu
Yugeng Liu
Ziqing Yang
Xinyue Shen
Michael Backes
Yang Zhang
AAML
35
66
0
08 Feb 2024
DeSparsify: Adversarial Attack Against Token Sparsification Mechanisms
  in Vision Transformers
DeSparsify: Adversarial Attack Against Token Sparsification Mechanisms in Vision Transformers
Oryan Yehezkel
Alon Zolfi
Amit Baras
Yuval Elovici
A. Shabtai
AAML
29
0
0
04 Feb 2024
Silent Guardian: Protecting Text from Malicious Exploitation by Large
  Language Models
Silent Guardian: Protecting Text from Malicious Exploitation by Large Language Models
Jiawei Zhao
Kejiang Chen
Xianjian Yuan
Yuang Qi
Weiming Zhang
Neng H. Yu
59
8
0
15 Dec 2023
Beyond Boundaries: A Comprehensive Survey of Transferable Attacks on AI Systems
Beyond Boundaries: A Comprehensive Survey of Transferable Attacks on AI Systems
Guangjing Wang
Ce Zhou
Yuanda Wang
Bocheng Chen
Hanqing Guo
Qiben Yan
AAML
SILM
60
3
0
20 Nov 2023
BERT Lost Patience Won't Be Robust to Adversarial Slowdown
BERT Lost Patience Won't Be Robust to Adversarial Slowdown
Zachary Coalson
Gabriel Ritter
Rakesh Bobba
Sanghyun Hong
AAML
19
1
0
29 Oct 2023
Invisible Threats: Backdoor Attack in OCR Systems
Invisible Threats: Backdoor Attack in OCR Systems
Mauro Conti
Nicola Farronato
Stefanos Koffas
Luca Pajola
S. Picek
32
1
0
12 Oct 2023
Jailbreak and Guard Aligned Language Models with Only Few In-Context
  Demonstrations
Jailbreak and Guard Aligned Language Models with Only Few In-Context Demonstrations
Zeming Wei
Yifei Wang
Ang Li
Yichuan Mo
Yisen Wang
42
235
0
10 Oct 2023
Enhancing Robust Representation in Adversarial Training: Alignment and
  Exclusion Criteria
Enhancing Robust Representation in Adversarial Training: Alignment and Exclusion Criteria
Nuoyan Zhou
Nannan Wang
Decheng Liu
Dawei Zhou
Xinbo Gao
AAML
28
2
0
05 Oct 2023
WASA: WAtermark-based Source Attribution for Large Language
  Model-Generated Data
WASA: WAtermark-based Source Attribution for Large Language Model-Generated Data
Jingtan Wang
Xinyang Lu
Zitong Zhao
Zhongxiang Dai
Chuan-Sheng Foo
See-Kiong Ng
K. H. Low
WaLM
57
14
0
01 Oct 2023
Unbiased Watermark for Large Language Models
Unbiased Watermark for Large Language Models
Zhengmian Hu
Lichang Chen
Xidong Wu
Yihan Wu
Hongyang R. Zhang
Heng-Chiao Huang
WaLM
38
45
0
22 Sep 2023
"Do Anything Now": Characterizing and Evaluating In-The-Wild Jailbreak
  Prompts on Large Language Models
"Do Anything Now": Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models
Xinyue Shen
Z. Chen
Michael Backes
Yun Shen
Yang Zhang
SILM
40
245
0
07 Aug 2023
LLM Censorship: A Machine Learning Challenge or a Computer Security
  Problem?
LLM Censorship: A Machine Learning Challenge or a Computer Security Problem?
David Glukhov
Ilia Shumailov
Y. Gal
Nicolas Papernot
V. Papyan
AAML
ELM
26
57
0
20 Jul 2023
Efficient Parallel Output-Sensitive Edit Distance
Efficient Parallel Output-Sensitive Edit Distance
Xiangyun Ding
Xiaojun Dong
Yan Gu
Youzhe Liu
Yihan Sun
15
4
0
30 Jun 2023
When Vision Fails: Text Attacks Against ViT and OCR
When Vision Fails: Text Attacks Against ViT and OCR
Nicholas Boucher
Jenny Blessing
Ilia Shumailov
Ross J. Anderson
Nicolas Papernot
AAML
26
4
0
12 Jun 2023
Enhancing Robustness of AI Offensive Code Generators via Data
  Augmentation
Enhancing Robustness of AI Offensive Code Generators via Data Augmentation
Cristina Improta
Pietro Liguori
R. Natella
B. Cukic
Domenico Cotroneo
AAML
30
2
0
08 Jun 2023
Avoid Adversarial Adaption in Federated Learning by Multi-Metric
  Investigations
Avoid Adversarial Adaption in Federated Learning by Multi-Metric Investigations
T. Krauß
Alexandra Dmitrienko
AAML
22
4
0
06 Jun 2023
Two-in-One: A Model Hijacking Attack Against Text Generation Models
Two-in-One: A Model Hijacking Attack Against Text Generation Models
Waiman Si
Michael Backes
Yang Zhang
A. Salem
SILM
19
22
0
12 May 2023
Boosting Big Brother: Attacking Search Engines with Encodings
Boosting Big Brother: Attacking Search Engines with Encodings
Nicholas Boucher
Luca Pajola
Ilia Shumailov
Ross J. Anderson
Mauro Conti
SILM
29
10
0
27 Apr 2023
RoCOCO: Robustness Benchmark of MS-COCO to Stress-test Image-Text
  Matching Models
RoCOCO: Robustness Benchmark of MS-COCO to Stress-test Image-Text Matching Models
Seulki Park
Daeho Um
Hajung Yoon
Sanghyuk Chun
Sangdoo Yun
Jin Young Choi
38
2
0
21 Apr 2023
RNN-Guard: Certified Robustness Against Multi-frame Attacks for
  Recurrent Neural Networks
RNN-Guard: Certified Robustness Against Multi-frame Attacks for Recurrent Neural Networks
Yunruo Zhang
Tianyu Du
S. Ji
Peng Tang
Shanqing Guo
AAML
26
2
0
17 Apr 2023
No more Reviewer #2: Subverting Automatic Paper-Reviewer Assignment
  using Adversarial Learning
No more Reviewer #2: Subverting Automatic Paper-Reviewer Assignment using Adversarial Learning
Thorsten Eisenhofer
Erwin Quiring
Jonas Moller
Doreen Riepel
Thorsten Holz
Konrad Rieck
AAML
21
6
0
25 Mar 2023
Model-tuning Via Prompts Makes NLP Models Adversarially Robust
Model-tuning Via Prompts Makes NLP Models Adversarially Robust
Mrigank Raman
Pratyush Maini
J. Zico Kolter
Zachary Chase Lipton
Danish Pruthi
AAML
25
17
0
13 Mar 2023
Learning the Legibility of Visual Text Perturbations
Learning the Legibility of Visual Text Perturbations
D. Seth
Rickard Stureborg
Danish Pruthi
Bhuwan Dhingra
AAML
41
4
0
09 Mar 2023
Not what you've signed up for: Compromising Real-World LLM-Integrated
  Applications with Indirect Prompt Injection
Not what you've signed up for: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection
Kai Greshake
Sahar Abdelnabi
Shailesh Mishra
C. Endres
Thorsten Holz
Mario Fritz
SILM
47
433
0
23 Feb 2023
MTTM: Metamorphic Testing for Textual Content Moderation Software
MTTM: Metamorphic Testing for Textual Content Moderation Software
Wenxuan Wang
Jen-tse Huang
Weibin Wu
Jianping Zhang
Yizhan Huang
Shuqing Li
Pinjia He
Michael Lyu
45
29
0
11 Feb 2023
Training-free Lexical Backdoor Attacks on Language Models
Training-free Lexical Backdoor Attacks on Language Models
Yujin Huang
Terry Yue Zhuo
Qiongkai Xu
Han Hu
Xingliang Yuan
Chunyang Chen
SILM
23
42
0
08 Feb 2023
On Robustness of Prompt-based Semantic Parsing with Large Pre-trained
  Language Model: An Empirical Study on Codex
On Robustness of Prompt-based Semantic Parsing with Large Pre-trained Language Model: An Empirical Study on Codex
Terry Yue Zhuo
Zhuang Li
Yujin Huang
Fatemeh Shiri
Weiqing Wang
Gholamreza Haffari
Yuan-Fang Li
AAML
23
53
0
30 Jan 2023
Demystifying Privacy Policy of Third-Party Libraries in Mobile Apps
Demystifying Privacy Policy of Third-Party Libraries in Mobile Apps
Kaifa Zhao
Xian Zhan
Le Yu
Shiyao Zhou
Hao Zhou
Xiapu Luo
Haoyu Wang
Yepang Liu
32
14
0
29 Jan 2023
A Watermark for Large Language Models
A Watermark for Large Language Models
John Kirchenbauer
Jonas Geiping
Yuxin Wen
Jonathan Katz
Ian Miers
Tom Goldstein
VLM
WaLM
13
460
0
24 Jan 2023
DDoD: Dual Denial of Decision Attacks on Human-AI Teams
DDoD: Dual Denial of Decision Attacks on Human-AI Teams
Benjamin Tag
N. V. Berkel
Sunny Verma
Benjamin Zi Hao Zhao
S. Berkovsky
Dali Kaafar
V. Kostakos
O. Ohrimenko
AAML
15
5
0
07 Dec 2022
On the Security Vulnerabilities of Text-to-SQL Models
On the Security Vulnerabilities of Text-to-SQL Models
Xutan Peng
Yipeng Zhang
Jingfeng Yang
Mark Stevenson
SILM
23
10
0
28 Nov 2022
Secure and Trustworthy Artificial Intelligence-Extended Reality (AI-XR)
  for Metaverses
Secure and Trustworthy Artificial Intelligence-Extended Reality (AI-XR) for Metaverses
Adnan Qayyum
M. A. Butt
Hassan Ali
Muhammad Usman
O. Halabi
Ala I. Al-Fuqaha
Q. Abbasi
Muhammad Ali Imran
Junaid Qadir
28
32
0
24 Oct 2022
ImpNet: Imperceptible and blackbox-undetectable backdoors in compiled
  neural networks
ImpNet: Imperceptible and blackbox-undetectable backdoors in compiled neural networks
Eleanor Clifford
Ilia Shumailov
Yiren Zhao
Ross J. Anderson
Robert D. Mullins
23
12
0
30 Sep 2022
Talking Trojan: Analyzing an Industry-Wide Disclosure
Talking Trojan: Analyzing an Industry-Wide Disclosure
Nicholas Boucher
Ross J. Anderson
19
3
0
22 Sep 2022
Exploiting Cultural Biases via Homoglyphs in Text-to-Image Synthesis
Exploiting Cultural Biases via Homoglyphs in Text-to-Image Synthesis
Lukas Struppek
Dominik Hintersdorf
Felix Friedrich
Manuel Brack
P. Schramowski
Kristian Kersting
73
26
0
19 Sep 2022
Fact-Saboteurs: A Taxonomy of Evidence Manipulation Attacks against
  Fact-Verification Systems
Fact-Saboteurs: A Taxonomy of Evidence Manipulation Attacks against Fact-Verification Systems
Sahar Abdelnabi
Mario Fritz
AAML
192
5
0
07 Sep 2022
12
Next