301

Scalable Lipschitz Residual Networks with Convex Potential Flows

International Conference on Machine Learning (ICML), 2021
Abstract

The Lipschitz constant of neural networks has been established as a key property to enforce the robustness of neural networks to adversarial examples. However, recent attempts to build 11-Lipschitz Neural Networks have all shown limitations and robustness have to be traded for accuracy and scalability or vice versa. In this work, we first show that using convex potentials in a residual network gradient flow provides a built-in 11-Lipschitz transformation. From this insight, we leverage the work on Input Convex Neural Networks to parametrize efficient layers with this property. A comprehensive set of experiments on CIFAR-10 demonstrates the scalability of our architecture and the benefit of our approach for 2\ell_2 provable defenses. Indeed, we train very deep and wide neural networks (up to 10001000 layers) and reach state-of-the-art results in terms of standard and certified accuracy, along with empirical robustness, in comparison with other 11-Lipschitz architectures.

View on arXiv
Comments on this paper