25
4

Manifold learning in Wasserstein space

Abstract

This paper aims at building the theoretical foundations for manifold learning algorithms in the space of absolutely continuous probability measures Pa.c.(Ω)\mathcal{P}_{\mathrm{a.c.}}(\Omega) with Ω\Omega a compact and convex subset of Rd\mathbb{R}^d, metrized with the Wasserstein-2 distance W\mathbb{W}. We begin by introducing a construction of submanifolds Λ\Lambda in Pa.c.(Ω)\mathcal{P}_{\mathrm{a.c.}}(\Omega) equipped with metric WΛ\mathbb{W}_\Lambda, the geodesic restriction of W\mathbb{W} to Λ\Lambda. In contrast to other constructions, these submanifolds are not necessarily flat, but still allow for local linearizations in a similar fashion to Riemannian submanifolds of Rd\mathbb{R}^d. We then show how the latent manifold structure of (Λ,WΛ)(\Lambda,\mathbb{W}_{\Lambda}) can be learned from samples {λi}i=1N\{\lambda_i\}_{i=1}^N of Λ\Lambda and pairwise extrinsic Wasserstein distances W\mathbb{W} on Pa.c.(Ω)\mathcal{P}_{\mathrm{a.c.}}(\Omega) only. In particular, we show that the metric space (Λ,WΛ)(\Lambda,\mathbb{W}_{\Lambda}) can be asymptotically recovered in the sense of Gromov--Wasserstein from a graph with nodes {λi}i=1N\{\lambda_i\}_{i=1}^N and edge weights W(λi,λj)W(\lambda_i,\lambda_j). In addition, we demonstrate how the tangent space at a sample λ\lambda can be asymptotically recovered via spectral analysis of a suitable ``covariance operator'' using optimal transport maps from λ\lambda to sufficiently close and diverse samples {λi}i=1N\{\lambda_i\}_{i=1}^N. The paper closes with some explicit constructions of submanifolds Λ\Lambda and numerical examples on the recovery of tangent spaces through spectral analysis.

View on arXiv
@article{hamm2025_2311.08549,
  title={ Manifold learning in Wasserstein space },
  author={ Keaton Hamm and Caroline Moosmüller and Bernhard Schmitzer and Matthew Thorpe },
  journal={arXiv preprint arXiv:2311.08549},
  year={ 2025 }
}
Comments on this paper