65
v1v2v3 (latest)

σσ-zero: Gradient-based Optimization of 0\ell_0-norm Adversarial Examples

Main:10 Pages
10 Figures
Bibliography:3 Pages
16 Tables
Appendix:10 Pages
Abstract

Evaluating the adversarial robustness of deep networks to gradient-based attacks is challenging. While most attacks consider 2\ell_2- and \ell_\infty-norm constraints to craft input perturbations, only a few investigate sparse 1\ell_1- and 0\ell_0-norm attacks. In particular, 0\ell_0-norm attacks remain the least studied due to the inherent complexity of optimizing over a non-convex and non-differentiable constraint. However, evaluating adversarial robustness under these attacks could reveal weaknesses otherwise left untested with more conventional 2\ell_2- and \ell_\infty-norm attacks. In this work, we propose a novel 0\ell_0-norm attack, called σ\sigma-zero, which leverages a differentiable approximation of the 0\ell_0 norm to facilitate gradient-based optimization, and an adaptive projection operator to dynamically adjust the trade-off between loss minimization and perturbation sparsity. Extensive evaluations using MNIST, CIFAR10, and ImageNet datasets, involving robust and non-robust models, show that σ\sigma\texttt{-zero} finds minimum 0\ell_0-norm adversarial examples without requiring any time-consuming hyperparameter tuning, and that it outperforms all competing sparse attacks in terms of success rate, perturbation size, and efficiency.

View on arXiv
Comments on this paper