ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2111.10754
11
0

Local Linearity and Double Descent in Catastrophic Overfitting

21 November 2021
Varun Sivashankar
Nikil Selvam
    AAML
ArXivPDFHTML
Abstract

Catastrophic overfitting is a phenomenon observed during Adversarial Training (AT) with the Fast Gradient Sign Method (FGSM) where the test robustness steeply declines over just one epoch in the training stage. Prior work has attributed this loss in robustness to a sharp decrease in local linearity\textit{local linearity}local linearity of the neural network with respect to the input space, and has demonstrated that introducing a local linearity measure as a regularization term prevents catastrophic overfitting. Using a simple neural network architecture, we experimentally demonstrate that maintaining high local linearity might be sufficient\textit{sufficient}sufficient to prevent catastrophic overfitting but is not necessary.\textit{necessary.}necessary. Further, inspired by Parseval networks, we introduce a regularization term to AT with FGSM to make the weight matrices of the network orthogonal and study the connection between orthogonality of the network weights and local linearity. Lastly, we identify the double descent\textit{double descent}double descent phenomenon during the adversarial training process.

View on arXiv
Comments on this paper