436

Adaptive Randomized Smoothing: Certifying Multi-Step Defences against Adversarial Examples

Neural Information Processing Systems (NeurIPS), 2024
Main:10 Pages
22 Figures
Bibliography:3 Pages
12 Tables
Appendix:12 Pages
Abstract

We propose Adaptive Randomized Smoothing (ARS) to certify the predictions of our test-time adaptive models against adversarial examples. ARS extends the analysis of randomized smoothing using f-Differential Privacy to certify the adaptive composition of multiple steps. For the first time, our theory covers the sound adaptive composition of general and high-dimensional functions of noisy input. We instantiate ARS on deep image classification to certify predictions against adversarial examples of bounded LL_{\infty} norm. In the LL_{\infty} threat model, our flexibility enables adaptation through high-dimensional input-dependent masking. We design adaptivity benchmarks, based on CIFAR-10 and CelebA, and show that ARS improves accuracy by 22 to 5%5\% points. On ImageNet, ARS improves accuracy by 11 to 3%3\% points over standard RS without adaptivity.

View on arXiv
Comments on this paper