
Rate distortion theory is concerned with optimally encoding a given signal class using a budget of bits, as . We say that can be compressed at rate if we can achieve an error of for encoding ; the supremal compression rate is denoted . Given a fixed coding scheme, there usually are elements of that are compressed at a higher rate than by the given coding scheme; we study the size of this set of signals. We show that for certain "nice" signal classes , a phase transition occurs: We construct a probability measure on such that for every coding scheme and any , the set of signals encoded with error by forms a -null-set. In particular our results apply to balls in Besov and Sobolev spaces that embed compactly into for a bounded Lipschitz domain . As an application, we show that several existing sharpness results concerning function approximation using deep neural networks are generically sharp. We also provide quantitative and non-asymptotic bounds on the probability that a random can be encoded to within accuracy using bits. This result is applied to the problem of approximately representing to within accuracy by a (quantized) neural network that is constrained to have at most nonzero weights and is generated by an arbitrary "learning" procedure. We show that for any there are constants such that, no matter how we choose the "learning" procedure, the probability of success is bounded from above by .
View on arXiv