Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1706.06083
Cited By
Towards Deep Learning Models Resistant to Adversarial Attacks
19 June 2017
A. Madry
Aleksandar Makelov
Ludwig Schmidt
Dimitris Tsipras
Adrian Vladu
SILM
OOD
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Towards Deep Learning Models Resistant to Adversarial Attacks"
50 / 6,484 papers shown
Title
Fast Adversarial Training against Sparse Attacks Requires Loss Smoothing
Xuyang Zhong
Yixiao Huang
Chen Liu
AAML
41
0
0
28 Feb 2025
Decoder Gradient Shield: Provable and High-Fidelity Prevention of Gradient-Based Box-Free Watermark Removal
Haonan An
Guang Hua
Zhengru Fang
Guowen Xu
Susanto Rahardja
Yuguang Fang
AAML
49
0
0
28 Feb 2025
Continuous Adversarial Text Representation Learning for Affective Recognition
Seungah Son
Andrez Saurez
Dongsoo Har
36
0
0
28 Feb 2025
QFAL: Quantum Federated Adversarial Learning
Walid El Maouaki
Nouhaila Innan
Alberto Marchisio
Taoufik Said
Mohamed Bennai
Muhammad Shafique
FedML
50
4
0
28 Feb 2025
Concealed Adversarial attacks on neural networks for sequential data
P. Sokerin
Dmitry Anikin
Sofia Krehova
Alexey Zaytsev
AAML
AI4TS
44
0
0
28 Feb 2025
Data-free Universal Adversarial Perturbation with Pseudo-semantic Prior
Chanhui Lee
Yeonghwan Song
Jeany Son
AAML
115
0
0
28 Feb 2025
UDora: A Unified Red Teaming Framework against LLM Agents by Dynamically Hijacking Their Own Reasoning
J. Zhang
Shuang Yang
B. Li
AAML
LLMAG
56
0
0
28 Feb 2025
Exploring the Impact of Temperature Scaling in Softmax for Classification and Adversarial Robustness
Hao Xuan
Bokai Yang
Xingyu Li
AAML
44
1
0
28 Feb 2025
SafeText: Safe Text-to-image Models via Aligning the Text Encoder
Yuepeng Hu
Zhengyuan Jiang
Neil Zhenqiang Gong
61
1
0
28 Feb 2025
HALO: Robust Out-of-Distribution Detection via Joint Optimisation
Hugo Lyons Keenan
S. Erfani
Christopher Leckie
OODD
199
0
0
27 Feb 2025
LISArD: Learning Image Similarity to Defend Against Gray-box Adversarial Attacks
Joana Cabral Costa
Tiago Roxo
Hugo Manuel Proença
Pedro R. M. Inácio
AAML
55
0
0
27 Feb 2025
Snowball Adversarial Attack on Traffic Sign Classification
Anthony Etim
Jakub Szefer
AAML
51
0
0
27 Feb 2025
Improving Adversarial Transferability in MLLMs via Dynamic Vision-Language Alignment Attack
Chenhe Gu
Jindong Gu
Andong Hua
Yao Qin
AAML
42
0
0
27 Feb 2025
Neural Antidote: Class-Wise Prompt Tuning for Purifying Backdoors in Pre-trained Vision-Language Models
Jiawei Kong
Hao Fang
Sihang Guo
Chenxi Qing
Bin Chen
Bin Wang
Shu-Tao Xia
AAML
VLM
90
0
0
26 Feb 2025
The NeRF Signature: Codebook-Aided Watermarking for Neural Radiance Fields
Ziyuan Luo
Anderson de Rezende Rocha
Boxin Shi
Qing Guo
Haoliang Li
Renjie Wan
47
0
0
26 Feb 2025
Model-Free Adversarial Purification via Coarse-To-Fine Tensor Network Representation
Guang Lin
D. Nguyen
Zerui Tao
Konstantinos Slavakis
Toshihisa Tanaka
Qibin Zhao
AAML
52
0
0
25 Feb 2025
CLIPure: Purification in Latent Space via CLIP for Adversarially Robust Zero-Shot Classification
Mingkun Zhang
Keping Bi
Wei Chen
J. Guo
Xueqi Cheng
BDL
VLM
50
1
0
25 Feb 2025
Improving the Transferability of Adversarial Examples by Inverse Knowledge Distillation
Wenyuan Wu
Zheng Liu
Yong Chen
Chao Su
Dezhong Peng
Xu Wang
AAML
34
0
0
24 Feb 2025
Interpreting Adversarial Attacks and Defences using Architectures with Enhanced Interpretability
Akshay G Rao
Chandrashekhar Lakshminarayanan
Arun Rajkumar
AI4CE
AAML
29
0
0
24 Feb 2025
Improved Diffusion-based Generative Model with Better Adversarial Robustness
Zekun Wang
Mingyang Yi
Shuchen Xue
Z. Li
Ming Liu
Bing Qin
Zhi-Ming Ma
DiffM
37
0
0
24 Feb 2025
MACPruning: Dynamic Operation Pruning to Mitigate Side-Channel DNN Model Extraction
Ruyi Ding
Cheng Gongye
Davis Ranney
A. A. Ding
Yunsi Fei
AAML
63
0
0
24 Feb 2025
SMTFL: Secure Model Training to Untrusted Participants in Federated Learning
Zhihui Zhao
Xiaorong Dong
Yimo Ren
Jianhua Wang
Dan Yu
Hongsong Zhu
Yongle Chen
77
0
0
24 Feb 2025
A stochastic smoothing framework for nonconvex-nonconcave min-sum-max problems with applications to Wasserstein distributionally robust optimization
Wei Liu
Muhammad Khan
Gabriel Mancino-Ball
Yangyang Xu
37
0
0
24 Feb 2025
EigenShield: Causal Subspace Filtering via Random Matrix Theory for Adversarially Robust Vision-Language Models
Nastaran Darabi
Devashri Naik
Sina Tayebati
Dinithi Jayasuriya
Ranganath Krishnan
A. R. Trivedi
AAML
41
0
0
24 Feb 2025
Unified Prompt Attack Against Text-to-Image Generation Models
Duo Peng
Qiuhong Ke
Mark He Huang
Ping Hu
J. Liu
43
0
0
23 Feb 2025
Class-Conditional Neural Polarizer: A Lightweight and Effective Backdoor Defense by Purifying Poisoned Features
Mingli Zhu
Shaokui Wei
Hongyuan Zha
Baoyuan Wu
AAML
37
0
0
23 Feb 2025
Tracking the Copyright of Large Vision-Language Models through Parameter Learning Adversarial Images
Yubo Wang
Jianting Tang
Chaohu Liu
Linli Xu
AAML
53
1
0
23 Feb 2025
Can Indirect Prompt Injection Attacks Be Detected and Removed?
Yulin Chen
Haoran Li
Yuan Sui
Yufei He
Yue Liu
Y. Song
Bryan Hooi
AAML
42
3
0
23 Feb 2025
A generative approach to LLM harmfulness detection with special red flag tokens
Sophie Xhonneux
David Dobre
Mehrnaz Mohfakhami
Leo Schwinn
Gauthier Gidel
47
1
0
22 Feb 2025
SEA: Shareable and Explainable Attribution for Query-based Black-box Attacks
Yue Gao
Ilia Shumailov
Kassem Fawaz
AAML
129
0
0
21 Feb 2025
Tight Clusters Make Specialized Experts
Stefan K. Nielsen
R. Teo
Laziz U. Abdullaev
Tan M. Nguyen
MoE
56
2
0
21 Feb 2025
Nearshore Underwater Target Detection Meets UAV-borne Hyperspectral Remote Sensing: A Novel Hybrid-level Contrastive Learning Framework and Benchmark Dataset
Jiahao Qi
Chuanhong Zhou
Xingyue Liu
Chen Chen
Dehui Zhu
Kangcheng Bin
Ping Zhong
74
0
0
21 Feb 2025
A Transfer Attack to Image Watermarks
Yuepeng Hu
Zhengyuan Jiang
Moyang Guo
Neil Zhenqiang Gong
77
10
0
20 Feb 2025
Carefully Blending Adversarial Training, Purification, and Aggregation Improves Adversarial Robustness
Emanuele Ballarin
A. Ansuini
Luca Bortolussi
AAML
62
0
0
20 Feb 2025
CyberSentinel: An Emergent Threat Detection System for AI Security
Krti Tallam
35
2
0
20 Feb 2025
Robust Optimization with Diffusion Models for Green Security
Lingkai Kong
Haichuan Wang
Yuqi Pan
Cheol Woo Kim
Mingxiao Song
Alayna Nguyen
Tonghan Wang
Haifeng Xu
Milind Tambe
40
0
0
19 Feb 2025
Artificial Kuramoto Oscillatory Neurons
Takeru Miyato
Sindy Lowe
Andreas Geiger
Max Welling
AI4CE
67
6
0
17 Feb 2025
Adversary-Aware DPO: Enhancing Safety Alignment in Vision Language Models via Adversarial Training
Fenghua Weng
Jian Lou
Jun Feng
Minlie Huang
Wenjie Wang
AAML
64
1
0
17 Feb 2025
PAR-AdvGAN: Improving Adversarial Attack Capability with Progressive Auto-Regression AdvGAN
Jiayu Zhang
Zhiyu Zhu
Xinyi Wang
Silin Liao
Zhibo Jin
Flora Salim
Huaming Chen
GAN
49
0
0
16 Feb 2025
FaceSwapGuard: Safeguarding Facial Privacy from DeepFake Threats through Identity Obfuscation
Li Wang
Zheng Li
X. Zhang
Shouling Ji
Shanqing Guo
PICV
AAML
56
1
0
15 Feb 2025
Wasserstein distributional adversarial training for deep neural networks
Xingjian Bai
Guangyi He
Yifan Jiang
Jan Obloj
OOD
56
0
0
13 Feb 2025
DeepSeek on a Trip: Inducing Targeted Visual Hallucinations via Representation Vulnerabilities
Chashi Mahiul Islam
Samuel Jacob Chacko
Preston Horne
Xiuwen Liu
102
0
0
11 Feb 2025
Robust Watermarks Leak: Channel-Aware Feature Extraction Enables Adversarial Watermark Manipulation
Zhongjie Ba
Y. Zhang
Peng Cheng
Bin Gong
Xinyu Zhang
Qinglong Wang
Kui Ren
AAML
89
0
0
10 Feb 2025
Amnesia as a Catalyst for Enhancing Black Box Pixel Attacks in Image Classification and Object Detection
Dongsu Song
Daehwa Ko
Jay Hoon Jung
AAML
57
0
0
10 Feb 2025
Effective Black-Box Multi-Faceted Attacks Breach Vision Large Language Model Guardrails
Yijun Yang
L. Wang
Xiao Yang
Lanqing Hong
Jun Zhu
AAML
61
0
0
09 Feb 2025
Sign-Symmetry Learning Rules are Robust Fine-Tuners
Aymene Berriche
Mehdi Zakaria Adjal
Riyadh Baghdadi
AAML
47
0
0
09 Feb 2025
Jailbreaking to Jailbreak
Jeremy Kritz
Vaughn Robinson
Robert Vacareanu
Bijan Varjavand
Michael Choi
Bobby Gogov
Scale Red Team
Summer Yue
Willow Primack
Zifan Wang
171
1
0
09 Feb 2025
Democratic Training Against Universal Adversarial Perturbations
Bing-Jie Sun
Jun Sun
Wei Zhao
AAML
57
0
0
08 Feb 2025
Adversarial Machine Learning: Attacks, Defenses, and Open Challenges
Pranav K Jha
AAML
51
0
0
08 Feb 2025
Confidence Elicitation: A New Attack Vector for Large Language Models
Brian Formento
Chuan-Sheng Foo
See-Kiong Ng
AAML
94
0
0
07 Feb 2025
Previous
1
2
3
4
5
6
...
128
129
130
Next