Dimension-agnostic inference using cross U-statistics
Classical asymptotic theory for statistical inference usually involves calibrating a statistic by fixing the dimension while letting the sample size increase to infinity. Recently, much effort has been dedicated towards understanding how these methods behave in high-dimensional settings, where and both increase to infinity together. This often leads to different inference procedures, depending on the assumptions about the dimensionality, leaving the practitioner in a bind: given a dataset with 100 samples in 20 dimensions, should they calibrate by assuming , or ? This paper considers the goal of dimension-agnostic inference; developing methods whose validity does not depend on any assumption on versus . We introduce an approach that uses variational representations of existing test statistics along with sample splitting and self-normalization to produce a new test statistic with a Gaussian limiting distribution. The resulting statistic can be viewed as a careful modification of degenerate U-statistics, dropping diagonal blocks and retaining off-diagonal blocks. We exemplify our technique for a handful of classical problems including one-sample mean and covariance testing. Our tests are shown to have minimax rate-optimal power against appropriate local alternatives, and their power is optimal up to a factor. We end by suggesting some next steps for extending dimension-agnostic inference to other problems.
View on arXiv