ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2110.03187
6
31

On the Optimal Memorization Power of ReLU Neural Networks

7 October 2021
Gal Vardi
Gilad Yehudai
Ohad Shamir
ArXivPDFHTML
Abstract

We study the memorization power of feedforward ReLU neural networks. We show that such networks can memorize any NNN points that satisfy a mild separability assumption using O~(N)\tilde{O}\left(\sqrt{N}\right)O~(N​) parameters. Known VC-dimension upper bounds imply that memorizing NNN samples requires Ω(N)\Omega(\sqrt{N})Ω(N​) parameters, and hence our construction is optimal up to logarithmic factors. We also give a generalized construction for networks with depth bounded by 1≤L≤N1 \leq L \leq \sqrt{N}1≤L≤N​, for memorizing NNN samples using O~(N/L)\tilde{O}(N/L)O~(N/L) parameters. This bound is also optimal up to logarithmic factors. Our construction uses weights with large bit complexity. We prove that having such a large bit complexity is both necessary and sufficient for memorization with a sub-linear number of parameters.

View on arXiv
Comments on this paper