ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2206.06496
16
4

Towards Alternative Techniques for Improving Adversarial Robustness: Analysis of Adversarial Training at a Spectrum of Perturbations

13 June 2022
Kaustubh Sridhar
Souradeep Dutta
Ramneet Kaur
James Weimer
O. Sokolsky
Insup Lee
    AAML
ArXivPDFHTML
Abstract

Adversarial training (AT) and its variants have spearheaded progress in improving neural network robustness to adversarial perturbations and common corruptions in the last few years. Algorithm design of AT and its variants are focused on training models at a specified perturbation strength ϵ\epsilonϵ and only using the feedback from the performance of that ϵ\epsilonϵ-robust model to improve the algorithm. In this work, we focus on models, trained on a spectrum of ϵ\epsilonϵ values. We analyze three perspectives: model performance, intermediate feature precision and convolution filter sensitivity. In each, we identify alternative improvements to AT that otherwise wouldn't have been apparent at a single ϵ\epsilonϵ. Specifically, we find that for a PGD attack at some strength δ\deltaδ, there is an AT model at some slightly larger strength ϵ\epsilonϵ, but no greater, that generalizes best to it. Hence, we propose overdesigning for robustness where we suggest training models at an ϵ\epsilonϵ just above δ\deltaδ. Second, we observe (across various ϵ\epsilonϵ values) that robustness is highly sensitive to the precision of intermediate features and particularly those after the first and second layer. Thus, we propose adding a simple quantization to defenses that improves accuracy on seen and unseen adaptive attacks. Third, we analyze convolution filters of each layer of models at increasing ϵ\epsilonϵ and notice that those of the first and second layer may be solely responsible for amplifying input perturbations. We present our findings and demonstrate our techniques through experiments with ResNet and WideResNet models on the CIFAR-10 and CIFAR-10-C datasets.

View on arXiv
Comments on this paper