277

Robust Neural Networks using Randomized Adversarial Training

Abstract

Since the discovery of adversarial examples in machine learning, researchers have designed several techniques to train neural networks that are robust against different types of attacks (most notably \ell_\infty and 2\ell_2 based attacks). However, it has been observed that the defense mechanisms designed to protect against one type of attack often offer poor performance against the other. In this paper, we introduce Randomized Adversarial Training (RAT), a technique that is efficient both against 2\ell_2 and \ell_\infty attacks. To obtain this result, we build upon adversarial training, a technique that is efficient against \ell_\infty attacks, and demonstrate that adding random noise at training and inference time further improves performance against \ltwo attacks. We then show that RAT is as efficient as adversarial training against \ell_\infty attacks while being robust against strong 2\ell_2 attacks. Our final comparative experiments demonstrate that RAT outperforms all state-of-the-art approaches against 2\ell_2 and \ell_\infty attacks.

View on arXiv
Comments on this paper