Universal Regular Conditional Distributions via Probability
Measure-Valued Deep Neural Models
This paper introduces a general framework for explicitly constructing universal deep neural models with inputs from a complete, separable, and locally-compact metric space and outputs in the Wasserstein-1 space over a complete and separable metric space . We find that any model built using the proposed framework is dense in the space of continuous functions from to in the corresponding uniform convergence on compacts topology, quantitatively. We identify two methods in which the curse of dimensionality can be broken. The first approach constructs subsets of consisting of functions that can be efficiently approximated. In the second approach, given any fixed , we build non-trivial subsets of on which can be efficiently approximated. The results are applied to three open problems lying at the interface of applied probability and computational learning theory. We find that the proposed models can approximate any regular conditional distribution of a -valued random element depending on an -valued random element , with arbitrarily high probability. The proposed models are also shown to be capable of generically expressing the aleatoric uncertainty present in most randomized machine learning models. The proposed framework is used to derive an affirmative answer to the open conjecture of Bishop (1994); namely: mixture density networks are generic regular conditional distributions. Numerical experiments are performed in the contexts of extreme learning machines, randomized DNNs, and heteroscedastic regression.
View on arXiv