Universal Multi-Party Poisoning Attacks

In this work, we demonstrate universal multi-party poisoning attacks that adapt and apply to any multi-party learning process with arbitrary interaction pattern between the parties. More generally, we introduce and study -poisoning attacks in which an adversary controls of the parties, and for each corrupted party , the adversary submits some poisoned data on behalf of that is still ``-close'' to the correct data (e.g., fraction of is still honestly generated). We prove that for any ``bad'' property of the final trained hypothesis (e.g., failing on a particular test example or having ``large'' risk) that has an arbitrarily small constant probability of happening without the attack, there always is a -poisoning attack that increases the probability of from to by . Our attack only uses clean labels, and it is online. More generally, we prove that for any bounded function defined over an -step random process , an adversary who can override each of the blocks with even dependent probability can increase the expected output by at least .
View on arXiv