13
79

Random Smoothing Might be Unable to Certify \ell_\infty Robustness for High-Dimensional Images

Abstract

We show a hardness result for random smoothing to achieve certified adversarial robustness against attacks in the p\ell_p ball of radius ϵ\epsilon when p>2p>2. Although random smoothing has been well understood for the 2\ell_2 case using the Gaussian distribution, much remains unknown concerning the existence of a noise distribution that works for the case of p>2p>2. This has been posed as an open problem by Cohen et al. (2019) and includes many significant paradigms such as the \ell_\infty threat model. In this work, we show that any noise distribution D\mathcal{D} over Rd\mathbb{R}^d that provides p\ell_p robustness for all base classifiers with p>2p>2 must satisfy Eηi2=Ω(d12/pϵ2(1δ)/δ2)\mathbb{E}\eta_i^2=\Omega(d^{1-2/p}\epsilon^2(1-\delta)/\delta^2) for 99% of the features (pixels) of vector ηD\eta\sim\mathcal{D}, where ϵ\epsilon is the robust radius and δ\delta is the score gap between the highest-scored class and the runner-up. Therefore, for high-dimensional images with pixel values bounded in [0,255][0,255], the required noise will eventually dominate the useful information in the images, leading to trivial smoothed classifiers.

View on arXiv
Comments on this paper