ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.21128
34
0

Squared families: Searching beyond regular probability models

27 March 2025
Russell Tsuchida
Jiawei Liu
Cheng Soon Ong
Dino Sejdinovic
ArXivPDFHTML
Abstract

We introduce squared families, which are families of probability densities obtained by squaring a linear transformation of a statistic. Squared families are singular, however their singularity can easily be handled so that they form regular models. After handling the singularity, squared families possess many convenient properties. Their Fisher information is a conformal transformation of the Hessian metric induced from a Bregman generator. The Bregman generator is the normalising constant, and yields a statistical divergence on the family. The normalising constant admits a helpful parameter-integral factorisation, meaning that only one parameter-independent integral needs to be computed for all normalising constants in the family, unlike in exponential families. Finally, the squared family kernel is the only integral that needs to be computed for the Fisher information, statistical divergence and normalising constant. We then describe how squared families are special in the broader class of ggg-families, which are obtained by applying a sufficiently regular function ggg to a linear transformation of a statistic. After removing special singularities, positively homogeneous families and exponential families are the only ggg-families for which the Fisher information is a conformal transformation of the Hessian metric, where the generator depends on the parameter only through the normalising constant. Even-order monomial families also admit parameter-integral factorisations, unlike exponential families. We study parameter estimation and density estimation in squared families, in the well-specified and misspecified settings. We use a universal approximation property to show that squared families can learn sufficiently well-behaved target densities at a rate of O(N−1/2)+Cn−1/4\mathcal{O}(N^{-1/2})+C n^{-1/4}O(N−1/2)+Cn−1/4, where NNN is the number of datapoints, nnn is the number of parameters, and CCC is some constant.

View on arXiv
@article{tsuchida2025_2503.21128,
  title={ Squared families: Searching beyond regular probability models },
  author={ Russell Tsuchida and Jiawei Liu and Cheng Soon Ong and Dino Sejdinovic },
  journal={arXiv preprint arXiv:2503.21128},
  year={ 2025 }
}
Comments on this paper