ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2105.07743
13
3

Universal Regular Conditional Distributions

17 May 2021
Anastasis Kratsios
ArXivPDFHTML
Abstract

We introduce a deep learning model that can universally approximate regular conditional distributions (RCDs). The proposed model operates in three phases: first, it linearizes inputs from a given metric space X\mathcal{X}X to Rd\mathbb{R}^dRd via a feature map, then a deep feedforward neural network processes these linearized features, and then the network's outputs are then transformed to the 111-Wasserstein space P1(RD)\mathcal{P}_1(\mathbb{R}^D)P1​(RD) via a probabilistic extension of the attention mechanism of Bahdanau et al.\ (2014). Our model, called the \textit{probabilistic transformer (PT)}, can approximate any continuous function from Rd\mathbb{R}^d Rd to P1(RD)\mathcal{P}_1(\mathbb{R}^D)P1​(RD) uniformly on compact sets, quantitatively. We identify two ways in which the PT avoids the curse of dimensionality when approximating P1(RD)\mathcal{P}_1(\mathbb{R}^D)P1​(RD)-valued functions. The first strategy builds functions in C(Rd,P1(RD))C(\mathbb{R}^d,\mathcal{P}_1(\mathbb{R}^D))C(Rd,P1​(RD)) which can be efficiently approximated by a PT, uniformly on any given compact subset of Rd\mathbb{R}^dRd. In the second approach, given any function fff in C(Rd,P1(RD))C(\mathbb{R}^d,\mathcal{P}_1(\mathbb{R}^D))C(Rd,P1​(RD)), we build compact subsets of Rd\mathbb{R}^dRd whereon fff can be efficiently approximated by a PT.

View on arXiv
Comments on this paper