ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1008.0526
220
152

Minimax risks for sparse regressions: Ultra-high-dimensional phenomenons

3 August 2010
Nicolas Verzélen
ArXivPDFHTML
Abstract

Consider the standard Gaussian linear regression model Y=Xθ+ϵY=X\theta+\epsilonY=Xθ+ϵ, where Y∈RnY\in R^nY∈Rn is a response vector and X∈Rn∗p X\in R^{n*p}X∈Rn∗p is a design matrix. Numerous work have been devoted to building efficient estimators of θ\thetaθ when ppp is much larger than nnn. In such a situation, a classical approach amounts to assume that θ0\theta_0θ0​ is approximately sparse. This paper studies the minimax risks of estimation and testing over classes of kkk-sparse vectors θ\thetaθ. These bounds shed light on the limitations due to high-dimensionality. The results encompass the problem of prediction (estimation of XθX\thetaXθ), the inverse problem (estimation of θ0\theta_0θ0​) and linear testing (testing Xθ=0X\theta=0Xθ=0). Interestingly, an elbow effect occurs when the number of variables klog⁡(p/k)k\log(p/k)klog(p/k) becomes large compared to nnn. Indeed, the minimax risks and hypothesis separation distances blow up in this ultra-high dimensional setting. We also prove that even dimension reduction techniques cannot provide satisfying results in an ultra-high dimensional setting. Moreover, we compute the minimax risks when the variance of the noise is unknown. The knowledge of this variance is shown to play a significant role in the optimal rates of estimation and testing. All these minimax bounds provide a characterization of statistical problems that are so difficult so that no procedure can provide satisfying results.

View on arXiv
Comments on this paper