Bayesian variable selection for high dimensional generalized linear models: convergence rates of the fitted densities

Bayesian variable selection has gained much empirical success recently in a variety of applications when the number of explanatory variables is possibly much larger than the sample size . For generalized linear models, if most of the 's have very small effects on the response , we show that it is possible to use Bayesian variable selection to reduce overfitting caused by the curse of dimensionality . In this approach a suitable prior can be used to choose a few out of the many 's to model , so that the posterior will propose probability densities that are ``often close'' to the true density in some sense. The closeness can be described by a Hellinger distance between and that scales at a power very close to , which is the ``finite-dimensional rate'' corresponding to a low-dimensional situation. These findings extend some recent work of Jiang [Technical Report 05-02 (2005) Dept. Statistics, Northwestern Univ.] on consistency of Bayesian variable selection for binary classification.
View on arXiv