12
43

Universal Multi-Party Poisoning Attacks

Abstract

In this work, we demonstrate universal multi-party poisoning attacks that adapt and apply to any multi-party learning process with arbitrary interaction pattern between the parties. More generally, we introduce and study (k,p)(k,p)-poisoning attacks in which an adversary controls k[m]k\in[m] of the parties, and for each corrupted party PiP_i, the adversary submits some poisoned data Ti\mathcal{T}'_i on behalf of PiP_i that is still ``(1p)(1-p)-close'' to the correct data Ti\mathcal{T}_i (e.g., 1p1-p fraction of Ti\mathcal{T}'_i is still honestly generated). We prove that for any ``bad'' property BB of the final trained hypothesis hh (e.g., hh failing on a particular test example or having ``large'' risk) that has an arbitrarily small constant probability of happening without the attack, there always is a (k,p)(k,p)-poisoning attack that increases the probability of BB from μ\mu to by μ1pk/m=μ+Ω(pk/m)\mu^{1-p \cdot k/m} = \mu + \Omega(p \cdot k/m). Our attack only uses clean labels, and it is online. More generally, we prove that for any bounded function f(x1,,xn)[0,1]f(x_1,\dots,x_n) \in [0,1] defined over an nn-step random process X=(x1,,xn)\mathbf{X} = (x_1,\dots,x_n), an adversary who can override each of the nn blocks with even dependent probability pp can increase the expected output by at least Ω(pVar[f(x)])\Omega(p \cdot \mathrm{Var}[f(\mathbf{x})]).

View on arXiv
Comments on this paper