Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2007.04596
Cited By
Learning Over-Parametrized Two-Layer ReLU Neural Networks beyond NTK
Annual Conference Computational Learning Theory (COLT), 2020
9 July 2020
Yuanzhi Li
Tengyu Ma
Hongyang R. Zhang
MLT
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Learning Over-Parametrized Two-Layer ReLU Neural Networks beyond NTK"
21 / 21 papers shown
Feature learning is decoupled from generalization in high capacity neural networks
Niclas Goring
Charles London
Abdurrahman Hadi Erturk
Chris Mingard
Yoonsoo Nam
Ard A. Louis
OOD
MLT
309
2
0
25 Jul 2025
How Learnable Grids Recover Fine Detail in Low Dimensions: A Neural Tangent Kernel Analysis of Multigrid Parametric Encodings
International Conference on Learning Representations (ICLR), 2025
Samuel Audia
Soheil Feizi
Matthias Zwicker
Dinesh Manocha
292
2
0
18 Apr 2025
SGD Finds then Tunes Features in Two-Layer Neural Networks with near-Optimal Sample Complexity: A Case Study in the XOR problem
International Conference on Learning Representations (ICLR), 2023
Margalit Glasgow
MLT
411
26
0
26 Sep 2023
Why Shallow Networks Struggle to Approximate and Learn High Frequencies
Shijun Zhang
Hongkai Zhao
Yimin Zhong
Haomin Zhou
330
7
0
29 Jun 2023
Over-Parameterization Exponentially Slows Down Gradient Descent for Learning a Single Neuron
Annual Conference Computational Learning Theory (COLT), 2023
Weihang Xu
S. Du
434
22
0
20 Feb 2023
Vision Transformers provably learn spatial structure
Neural Information Processing Systems (NeurIPS), 2022
Samy Jelassi
Michael E. Sander
Yuan-Fang Li
ViT
MLT
298
103
0
13 Oct 2022
Neural Networks can Learn Representations with Gradient Descent
Annual Conference Computational Learning Theory (COLT), 2022
Alexandru Damian
Jason D. Lee
Mahdi Soltanolkotabi
SSL
MLT
381
175
0
30 Jun 2022
The Mechanism of Prediction Head in Non-contrastive Self-supervised Learning
Neural Information Processing Systems (NeurIPS), 2022
Zixin Wen
Yuanzhi Li
SSL
479
42
0
12 May 2022
Efficiently Learning Any One Hidden Layer ReLU Network From Queries
Sitan Chen
Adam R. Klivans
Raghu Meka
MLAU
MLT
343
9
0
08 Nov 2021
On the Provable Generalization of Recurrent Neural Networks
Lifu Wang
Bo Shen
Bo Hu
Xing Cao
438
9
0
29 Sep 2021
Deep Networks Provably Classify Data on Curves
Neural Information Processing Systems (NeurIPS), 2021
Tingran Wang
Sam Buchanan
D. Gilboa
John N. Wright
288
9
0
29 Jul 2021
Small random initialization is akin to spectral learning: Optimization and generalization guarantees for overparameterized low-rank matrix reconstruction
Neural Information Processing Systems (NeurIPS), 2021
Dominik Stöger
Mahdi Soltanolkotabi
ODL
500
92
0
28 Jun 2021
Proxy Convexity: A Unified Framework for the Analysis of Neural Networks Trained by Gradient Descent
Neural Information Processing Systems (NeurIPS), 2021
Spencer Frei
Quanquan Gu
396
29
0
25 Jun 2021
Toward Understanding the Feature Learning Process of Self-supervised Contrastive Learning
International Conference on Machine Learning (ICML), 2021
Zixin Wen
Yuanzhi Li
SSL
MLT
491
161
0
31 May 2021
Why Do Local Methods Solve Nonconvex Problems?
Tengyu Ma
172
14
0
24 Mar 2021
Unintended Effects on Adaptive Learning Rate for Training Neural Network with Output Scale Change
Ryuichi Kanoh
M. Sugiyama
209
0
0
05 Mar 2021
A Local Convergence Theory for Mildly Over-Parameterized Two-Layer Neural Network
Annual Conference Computational Learning Theory (COLT), 2021
Mo Zhou
Rong Ge
Chi Jin
395
53
0
04 Feb 2021
Provable Generalization of SGD-trained Neural Networks of Any Width in the Presence of Adversarial Label Noise
International Conference on Machine Learning (ICML), 2021
Spencer Frei
Yuan Cao
Quanquan Gu
FedML
MLT
533
22
0
04 Jan 2021
Towards Understanding Ensemble, Knowledge Distillation and Self-Distillation in Deep Learning
International Conference on Learning Representations (ICLR), 2020
Zeyuan Allen-Zhu
Yuanzhi Li
FedML
752
454
0
17 Dec 2020
A Modular Analysis of Provable Acceleration via Polyak's Momentum: Training a Wide ReLU Network and a Deep Linear Network
International Conference on Machine Learning (ICML), 2020
Jun-Kun Wang
Chi-Heng Lin
Jacob D. Abernethy
726
26
0
04 Oct 2020
Feature Purification: How Adversarial Training Performs Robust Deep Learning
Zeyuan Allen-Zhu
Yuanzhi Li
MLT
AAML
525
174
0
20 May 2020
1
Page 1 of 1