Coresets For Monotonic Functions with Applications to Deep Learning
- CLL
Coreset (or core-set) in this paper is a small weighted \emph{subset} of the input set with respect to a given \emph{monotonic} function that \emph{provably} approximates its fitting loss to \emph{any} given . Using we can obtain approximation to that minimizes this loss, by running \emph{existing} optimization algorithms on . We provide: (i) a lower bound that proves that there are sets with no coresets smaller than , (ii) a proof that a small coreset of size near-logarithmic in exists for \emph{any} input , under natural assumption that holds e.g. for logistic regression and the sigmoid activation function. (iii) a generic algorithm that computes in expected time, (iv) novel technique for improving existing deep networks using such coresets, (v) extensive experimental results with open code.oving existing deep networks using such coresets, (v) extensive experimental results with open code.
View on arXiv