14
8

Adversarial Parameter Attack on Deep Neural Networks

Abstract

In this paper, a new parameter perturbation attack on DNNs, called adversarial parameter attack, is proposed, in which small perturbations to the parameters of the DNN are made such that the accuracy of the attacked DNN does not decrease much, but its robustness becomes much lower. The adversarial parameter attack is stronger than previous parameter perturbation attacks in that the attack is more difficult to be recognized by users and the attacked DNN gives a wrong label for any modified sample input with high probability. The existence of adversarial parameters is proved. For a DNN FΘF_{\Theta} with the parameter set Θ\Theta satisfying certain conditions, it is shown that if the depth of the DNN is sufficiently large, then there exists an adversarial parameter set Θa\Theta_a for Θ\Theta such that the accuracy of FΘaF_{\Theta_a} is equal to that of FΘF_{\Theta}, but the robustness measure of FΘaF_{\Theta_a} is smaller than any given bound. An effective training algorithm is given to compute adversarial parameters and numerical experiments are used to demonstrate that the algorithms are effective to produce high quality adversarial parameters.

View on arXiv
Comments on this paper