-zero: Gradient-based Optimization of -norm Adversarial Examples
Evaluating the adversarial robustness of deep networks to gradient-based attacks is challenging. While most attacks consider - and -norm constraints to craft input perturbations, only a few investigate sparse - and -norm attacks. In particular, -norm attacks remain the least studied due to the inherent complexity of optimizing over a non-convex and non-differentiable constraint. However, evaluating adversarial robustness under these attacks could reveal weaknesses otherwise left untested with more conventional - and -norm attacks. In this work, we propose a novel -norm attack, called -zero, which leverages a differentiable approximation of the norm to facilitate gradient-based optimization, and an adaptive projection operator to dynamically adjust the trade-off between loss minimization and perturbation sparsity. Extensive evaluations using MNIST, CIFAR10, and ImageNet datasets, involving robust and non-robust models, show that \texttt{-zero} finds minimum -norm adversarial examples without requiring any time-consuming hyperparameter tuning, and that it outperforms all competing sparse attacks in terms of success rate, perturbation size, and efficiency.
View on arXiv