ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2407.05622
16
6

On the Complexity of Learning Sparse Functions with Statistical and Gradient Queries

8 July 2024
Nirmit Joshi
Theodor Misiakiewicz
Nathan Srebro
ArXivPDFHTML
Abstract

The goal of this paper is to investigate the complexity of gradient algorithms when learning sparse functions (juntas). We introduce a type of Statistical Queries (SQ\mathsf{SQ}SQ), which we call Differentiable Learning Queries (DLQ\mathsf{DLQ}DLQ), to model gradient queries on a specified loss with respect to an arbitrary model. We provide a tight characterization of the query complexity of DLQ\mathsf{DLQ}DLQ for learning the support of a sparse function over generic product distributions. This complexity crucially depends on the loss function. For the squared loss, DLQ\mathsf{DLQ}DLQ matches the complexity of Correlation Statistical Queries (CSQ)(\mathsf{CSQ})(CSQ)--potentially much worse than SQ\mathsf{SQ}SQ. But for other simple loss functions, including the ℓ1\ell_1ℓ1​ loss, DLQ\mathsf{DLQ}DLQ always achieves the same complexity as SQ\mathsf{SQ}SQ. We also provide evidence that DLQ\mathsf{DLQ}DLQ can indeed capture learning with (stochastic) gradient descent by showing it correctly describes the complexity of learning with a two-layer neural network in the mean field regime and linear scaling.

View on arXiv
Comments on this paper