Sponge examples are test-time inputs optimized to increase energy consumption and prediction latency of deep networks deployed on hardware accelerators. By increasing the fraction of neurons activated during classification, these attacks reduce sparsity in network activation patterns, worsening the performance of hardware accelerators. In this work, we present a novel training-time attack, named sponge poisoning, which aims to worsen energy consumption and prediction latency of neural networks on any test input without affecting classification accuracy. To stage this attack, we assume that the attacker can control only a few model updates during training -- a likely scenario, e.g., when model training is outsourced to an untrusted third party or distributed via federated learning. Our extensive experiments on image classification tasks show that sponge poisoning is effective, and that fine-tuning poisoned models to repair them poses prohibitive costs for most users, highlighting that tackling sponge poisoning remains an open issue.
View on arXiv@article{cinà2025_2203.08147, title={ Energy-Latency Attacks via Sponge Poisoning }, author={ Antonio Emanuele Cinà and Ambra Demontis and Battista Biggio and Fabio Roli and Marcello Pelillo }, journal={arXiv preprint arXiv:2203.08147}, year={ 2025 } }