ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1810.12437
53
20
v1v2v3v4v5v6 (latest)

Prior-preconditioned conjugate gradient method for accelerated Gibbs sampling in "large nnn & large ppp" Bayesian sparse regression

29 October 2018
A. Nishimura
M. Suchard
ArXiv (abs)PDFHTML
Abstract

In a modern observational study based on healthcare databases, the number of observations and of predictors typically range in the order of 10510^5105 ~ 10610^6106 and of 10410^4104 ~ 10510^5105. Despite the large sample size, data rarely provide sufficient information to reliably estimate such a large number of parameters. Sparse regression techniques provide potential solutions, one notable approach being the Bayesian methods based on shrinkage priors. In the "large n & large p" setting, however, posterior computation encounters a major bottleneck at repeated sampling from a high-dimensional Gaussian distribution, whose precision matrix Φ\PhiΦ is expensive to compute and factorize. In this article, we present a novel algorithm to speed up this bottleneck based on the following observation: we can cheaply generate a random vector bbb such that the solution to the linear system Φβ=b\Phi \beta = bΦβ=b has the desired Gaussian distribution. We can then solve the linear system by the conjugate gradient (CG) algorithm through matrix-vector multiplications by Φ\PhiΦ; this involves no explicit factorization or calculation of \bPhi\bPhi\bPhi itself. Rapid convergence of CG in this context is guaranteed by the theory of prior-preconditioning we develop. We apply our algorithm to a clinically relevant large-scale observational study with n = 72,489 patients and p = 22,175 clinical covariates, designed to assess the relative risk of adverse events from two alternative blood anti-coagulants. Our algorithm demonstrates an order of magnitude speed-up in the posterior computation.

View on arXiv
Comments on this paper