ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2006.02855
  4. Cited By
Network size and weights size for memorization with two-layers neural
  networks
v1v2 (latest)

Network size and weights size for memorization with two-layers neural networks

4 June 2020
Sébastien Bubeck
Ronen Eldan
Y. Lee
Dan Mikulincer
ArXiv (abs)PDFHTML

Papers citing "Network size and weights size for memorization with two-layers neural networks"

21 / 21 papers shown
Title
Analysis of the expected $L_2$ error of an over-parametrized deep neural
  network estimate learned by gradient descent without regularization
Analysis of the expected L2L_2L2​ error of an over-parametrized deep neural network estimate learned by gradient descent without regularization
Selina Drews
Michael Kohler
63
4
0
24 Nov 2023
What do larger image classifiers memorise?
What do larger image classifiers memorise?
Michal Lukasik
Vaishnavh Nagarajan
A. S. Rawat
A. Menon
Sanjiv Kumar
93
5
0
09 Oct 2023
Memorization with neural nets: going beyond the worst case
Memorization with neural nets: going beyond the worst case
S. Dirksen
Patrick Finke
Martin Genzel
82
0
0
30 Sep 2023
Memorization Capacity of Multi-Head Attention in Transformers
Memorization Capacity of Multi-Head Attention in Transformers
Sadegh Mahdavi
Renjie Liao
Christos Thrampoulidis
97
25
0
03 Jun 2023
Finite Sample Identification of Wide Shallow Neural Networks with Biases
Finite Sample Identification of Wide Shallow Neural Networks with Biases
M. Fornasier
T. Klock
Marco Mondelli
Michael Rauchensteiner
52
6
0
08 Nov 2022
When Expressivity Meets Trainability: Fewer than $n$ Neurons Can Work
When Expressivity Meets Trainability: Fewer than nnn Neurons Can Work
Jiawei Zhang
Yushun Zhang
Mingyi Hong
Ruoyu Sun
Zhi-Quan Luo
124
10
0
21 Oct 2022
Why Robust Generalization in Deep Learning is Difficult: Perspective of
  Expressive Power
Why Robust Generalization in Deep Learning is Difficult: Perspective of Expressive Power
Binghui Li
Jikai Jin
Han Zhong
John E. Hopcroft
Liwei Wang
OOD
118
27
0
27 May 2022
Memorization and Optimization in Deep Neural Networks with Minimum
  Over-parameterization
Memorization and Optimization in Deep Neural Networks with Minimum Over-parameterization
Simone Bombari
Mohammad Hossein Amani
Marco Mondelli
85
26
0
20 May 2022
On the Optimal Memorization Power of ReLU Neural Networks
On the Optimal Memorization Power of ReLU Neural Networks
Gal Vardi
Gilad Yehudai
Ohad Shamir
76
32
0
07 Oct 2021
NeuFENet: Neural Finite Element Solutions with Theoretical Bounds for
  Parametric PDEs
NeuFENet: Neural Finite Element Solutions with Theoretical Bounds for Parametric PDEs
Biswajit Khara
Aditya Balu
Ameya Joshi
Soumik Sarkar
Chinmay Hegde
A. Krishnamurthy
Baskar Ganapathysubramanian
106
20
0
04 Oct 2021
An Exponential Improvement on the Memorization Capacity of Deep
  Threshold Networks
An Exponential Improvement on the Memorization Capacity of Deep Threshold Networks
Shashank Rajput
Kartik K. Sreenivasan
Dimitris Papailiopoulos
Amin Karbasi
40
25
0
14 Jun 2021
A Recipe for Global Convergence Guarantee in Deep Neural Networks
A Recipe for Global Convergence Guarantee in Deep Neural Networks
Kenji Kawaguchi
Qingyun Sun
58
12
0
12 Apr 2021
When Are Solutions Connected in Deep Networks?
When Are Solutions Connected in Deep Networks?
Quynh N. Nguyen
Pierre Bréchet
Marco Mondelli
73
10
0
18 Feb 2021
Knowledge Infused Policy Gradients for Adaptive Pandemic Control
Knowledge Infused Policy Gradients for Adaptive Pandemic Control
Kaushik Roy
Qi Zhang
Manas Gaur
A. Sheth
48
12
0
11 Feb 2021
Neural Networks with Complex-Valued Weights Have No Spurious Local
  Minima
Neural Networks with Complex-Valued Weights Have No Spurious Local Minima
Xingtu Liu
MLT
41
0
0
31 Jan 2021
Tight Bounds on the Smallest Eigenvalue of the Neural Tangent Kernel for
  Deep ReLU Networks
Tight Bounds on the Smallest Eigenvalue of the Neural Tangent Kernel for Deep ReLU Networks
Quynh N. Nguyen
Marco Mondelli
Guido Montúfar
87
83
0
21 Dec 2020
A law of robustness for two-layers neural networks
A law of robustness for two-layers neural networks
Sébastien Bubeck
Yuanzhi Li
Dheeraj M. Nagaraj
83
57
0
30 Sep 2020
The Interpolation Phase Transition in Neural Networks: Memorization and
  Generalization under Lazy Training
The Interpolation Phase Transition in Neural Networks: Memorization and Generalization under Lazy Training
Andrea Montanari
Yiqiao Zhong
187
97
0
25 Jul 2020
Training (Overparametrized) Neural Networks in Near-Linear Time
Training (Overparametrized) Neural Networks in Near-Linear Time
Jan van den Brand
Binghui Peng
Zhao Song
Omri Weinstein
ODL
91
83
0
20 Jun 2020
On Sparsity in Overparametrised Shallow ReLU Networks
On Sparsity in Overparametrised Shallow ReLU Networks
Jaume de Dios
Joan Bruna
63
14
0
18 Jun 2020
A Corrective View of Neural Networks: Representation, Memorization and
  Learning
A Corrective View of Neural Networks: Representation, Memorization and Learning
Guy Bresler
Dheeraj M. Nagaraj
MLT
77
18
0
01 Feb 2020
1