445

Metric Hypertransformers are Universal Adapted Maps

Mathematical Finance (Math. Finance), 2022
Abstract

We introduce a universal class of geometric deep learning models, called metric hypertransformers (MHTs), capable of approximating any adapted map F:XZYZF:\mathscr{X}^{\mathbb{Z}}\rightarrow \mathscr{Y}^{\mathbb{Z}} with approximable complexity, where XRd\mathscr{X}\subseteq \mathbb{R}^d and Y\mathscr{Y} is any suitable metric space, and XZ\mathscr{X}^{\mathbb{Z}} (resp. YZ\mathscr{Y}^{\mathbb{Z}}) capture all discrete-time paths on X\mathscr{X} (resp. Y\mathscr{Y}). Suitable spaces Y\mathscr{Y} include various (adapted) Wasserstein spaces, all Fr\'{e}chet spaces admitting a Schauder basis, and a variety of Riemannian manifolds arising from information geometry. Even in the static case, where f:XYf:\mathscr{X}\rightarrow \mathscr{Y} is a H\"{o}lder map, our results provide the first (quantitative) universal approximation theorem compatible with any such X\mathscr{X} and Y\mathscr{Y}. Our universal approximation theorems are quantitative, and they depend on the regularity of FF, the choice of activation function, the metric entropy and diameter of X\mathscr{X}, and on the regularity of the compact set of paths whereon the approximation is performed. Our guiding examples originate from mathematical finance. Notably, the MHT models introduced here are able to approximate a broad range of stochastic processes' kernels, including solutions to SDEs, many processes with arbitrarily long memory, and functions mapping sequential data to sequences of forward rate curves.

View on arXiv
Comments on this paper