ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1810.01323
41
5
v1v2 (latest)

Moderate-Dimensional Inferences on Quadratic Functionals in Ordinary Least Squares

2 October 2018
Xiao Guo
Guang Cheng
ArXiv (abs)PDFHTML
Abstract

Statistical inferences for quadratic functionals of linear regression parameter have found wide applications including signal detection, global testing, inferences of error variance and fraction of variance explained. Classical theory based on ordinary least squares estimator works perfectly in the low-dimensional regime, but fails when the parameter dimension pnp_npn​ grows proportionally to the sample size nnn. In some cases, its performance is not satisfactory even when n≥5pnn\ge 5p_nn≥5pn​. The main contribution of this paper is to develop {\em dimension-adaptive} inferences for quadratic functionals when lim⁡n→∞pn/n=τ∈[0,1)\lim_{n\to \infty} p_n/n=\tau\in[0,1)limn→∞​pn​/n=τ∈[0,1). We propose a bias-and-variance-corrected test statistic and demonstrate that its theoretical validity (such as consistency and asymptotic normality) is adaptive to both low dimension with τ=0\tau = 0τ=0 and moderate dimension with τ∈(0,1)\tau \in(0, 1)τ∈(0,1). Our general theory holds, in particular, without Gaussian design/error or structural parameter assumption, and applies to a broad class of quadratic functionals covering all aforementioned applications. As a by-product, we find that the classical fixed-dimensional results continue to hold {\em if and only if} the signal-to-noise ratio is large enough, say when pnp_npn​ diverges but slower than nnn. Extensive numerical results demonstrate the satisfactory performance of the proposed methodology even when pn≥0.9np_n\ge 0.9npn​≥0.9n in some extreme cases. The mathematical arguments are based on the random matrix theory and leave-one-observation-out method.

View on arXiv
Comments on this paper