ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2102.04074
117
38

Learning Curve Theory

8 February 2021
Marcus Hutter
ArXivPDFHTML
Abstract

Recently a number of empirical "universal" scaling law papers have been published, most notably by OpenAI. `Scaling laws' refers to power-law decreases of training or test error w.r.t. more data, larger neural networks, and/or more compute. In this work we focus on scaling w.r.t. data size nnn. Theoretical understanding of this phenomenon is largely lacking, except in finite-dimensional models for which error typically decreases with n−1/2n^{-1/2}n−1/2 or n−1n^{-1}n−1, where nnn is the sample size. We develop and theoretically analyse the simplest possible (toy) model that can exhibit n−βn^{-\beta}n−β learning curves for arbitrary power β>0\beta>0β>0, and determine whether power laws are universal or depend on the data distribution.

View on arXiv
Comments on this paper