79
v1v2 (latest)

Accelerating Large-Scale Dataset Distillation via Exploration-Exploitation Optimization

Muhammad J. Alahmadi
Peng Gao
Feiyi Wang
Dongkuan Xu
Main:8 Pages
9 Figures
Bibliography:2 Pages
15 Tables
Appendix:5 Pages
Abstract

Dataset distillation compresses the original data into compact synthetic datasets, reducing training time and storage while retaining model performance, enabling deployment under limited resources. Although recent decoupling-based distillation methods enable dataset distillation at large scale, they continue to face an efficiency gap: optimization-based decoupling methods achieve higher accuracy but demand intensive computation, whereas optimization-free decoupling methods are efficient but sacrifice accuracy. To overcome this trade-off, we propose Exploration--Exploitation Distillation (E2^2D), a simple, practical method that minimizes redundant computation through an efficient pipeline that begins with full-image initialization to preserve semantic integrity and feature diversity. It then uses a two-phase optimization strategy: an exploration phase that performs uniform updates and identifies high-loss regions, and an exploitation phase that focuses updates on these regions to accelerate convergence. We evaluate E2^2D on large-scale benchmarks, surpassing the state-of-the-art on ImageNet-1K while being 18×18\times faster, and on ImageNet-21K, our method substantially improves accuracy while remaining 4.3×4.3\times faster. These results demonstrate that targeted, redundancy-reducing updates, rather than brute-force optimization, bridge the gap between accuracy and efficiency in large-scale dataset distillation. Code is available atthis https URL.

View on arXiv
Comments on this paper