Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1710.03667
Cited By
High-dimensional dynamics of generalization error in neural networks
10 October 2017
Madhu S. Advani
Andrew M. Saxe
AI4CE
Re-assign community
ArXiv
PDF
HTML
Papers citing
"High-dimensional dynamics of generalization error in neural networks"
50 / 296 papers shown
Title
On the generalization of learning algorithms that do not converge
N. Chandramoorthy
Andreas Loukas
Khashayar Gatmiry
Stefanie Jegelka
MLT
14
11
0
16 Aug 2022
Benign, Tempered, or Catastrophic: A Taxonomy of Overfitting
Neil Rohit Mallinar
James B. Simon
Amirhesam Abedsoltan
Parthe Pandit
M. Belkin
Preetum Nakkiran
24
37
0
14 Jul 2022
Studying Generalization Through Data Averaging
C. Gomez-Uribe
FedML
19
0
0
28 Jun 2022
How You Start Matters for Generalization
Sameera Ramasinghe
L. MacDonald
M. Farazi
Hemanth Saratchandran
Simon Lucey
ODL
AI4CE
31
6
0
17 Jun 2022
Trajectory-dependent Generalization Bounds for Deep Neural Networks via Fractional Brownian Motion
Chengli Tan
Jiang Zhang
Junmin Liu
35
1
0
09 Jun 2022
How does overparametrization affect performance on minority groups?
Subha Maity
Saptarshi Roy
Songkai Xue
Mikhail Yurochkin
Yuekai Sun
19
3
0
07 Jun 2022
Few-Shot Learning by Dimensionality Reduction in Gradient Space
M. Gauch
M. Beck
Thomas Adler
D. Kotsur
Stefan Fiel
...
Markus Holzleitner
Werner Zellinger
D. Klotz
Sepp Hochreiter
Sebastian Lehner
33
9
0
07 Jun 2022
Beyond accuracy: generalization properties of bio-plausible temporal credit assignment rules
Yuhan Helena Liu
Arna Ghosh
Blake A. Richards
E. Shea-Brown
Guillaume Lajoie
21
9
0
02 Jun 2022
Precise Learning Curves and Higher-Order Scaling Limits for Dot Product Kernel Regression
Lechao Xiao
Hong Hu
Theodor Misiakiewicz
Yue M. Lu
Jeffrey Pennington
59
18
0
30 May 2022
Scaling Laws and Interpretability of Learning from Repeated Data
Danny Hernandez
Tom B. Brown
Tom Conerly
Nova Dassarma
Dawn Drain
...
Catherine Olsson
Dario Amodei
Nicholas Joseph
Jared Kaplan
Sam McCandlish
18
108
0
21 May 2022
Self-Consistent Dynamical Field Theory of Kernel Evolution in Wide Neural Networks
Blake Bordelon
C. Pehlevan
MLT
24
79
0
19 May 2022
Ridgeless Regression with Random Features
Jian Li
Yong-Jin Liu
Yingying Zhang
19
2
0
01 May 2022
Overparameterized Linear Regression under Adversarial Attacks
Antônio H. Ribeiro
Thomas B. Schon
AAML
12
18
0
13 Apr 2022
Can Neural Nets Learn the Same Model Twice? Investigating Reproducibility and Double Descent from the Decision Boundary Perspective
Gowthami Somepalli
Liam H. Fowl
Arpit Bansal
Ping Yeh-Chiang
Yehuda Dar
Richard Baraniuk
Micah Goldblum
Tom Goldstein
11
64
0
15 Mar 2022
Phenomenology of Double Descent in Finite-Width Neural Networks
Sidak Pal Singh
Aurélien Lucchi
Thomas Hofmann
Bernhard Schölkopf
13
10
0
14 Mar 2022
Bias-variance decomposition of overparameterized regression with random linear features
J. Rocks
Pankaj Mehta
15
12
0
10 Mar 2022
Contrasting random and learned features in deep Bayesian linear regression
Jacob A. Zavatone-Veth
William L. Tong
C. Pehlevan
BDL
MLT
28
26
0
01 Mar 2022
On the Origins of the Block Structure Phenomenon in Neural Network Representations
Thao Nguyen
M. Raghu
Simon Kornblith
25
14
0
15 Feb 2022
Towards Data-Algorithm Dependent Generalization: a Case Study on Overparameterized Linear Regression
Jing Xu
Jiaye Teng
Yang Yuan
Andrew Chi-Chih Yao
21
1
0
12 Feb 2022
Investigating Power laws in Deep Representation Learning
Arna Ghosh
Arnab Kumar Mondal
Kumar Krishna Agrawal
Blake A. Richards
SSL
OOD
11
19
0
11 Feb 2022
Data-driven emergence of convolutional structure in neural networks
Alessandro Ingrosso
Sebastian Goldt
48
38
0
01 Feb 2022
Fluctuations, Bias, Variance & Ensemble of Learners: Exact Asymptotics for Convex Losses in High-Dimension
Bruno Loureiro
Cédric Gerbelot
Maria Refinetti
G. Sicuro
Florent Krzakala
18
25
0
31 Jan 2022
Error Scaling Laws for Kernel Classification under Source and Capacity Conditions
Hugo Cui
Bruno Loureiro
Florent Krzakala
Lenka Zdeborová
43
10
0
29 Jan 2022
A phase transition for finding needles in nonlinear haystacks with LASSO artificial neural networks
Xiaoyu Ma
S. Sardy
N. Hengartner
Nikolai Bobenko
Yen Ting Lin
24
2
0
21 Jan 2022
Largest Eigenvalues of the Conjugate Kernel of Single-Layered Neural Networks
L. Benigni
Sandrine Péché
20
8
0
13 Jan 2022
Complexity from Adaptive-Symmetries Breaking: Global Minima in the Statistical Mechanics of Deep Neural Networks
Shaun Li
AI4CE
30
0
0
03 Jan 2022
Multi-scale Feature Learning Dynamics: Insights for Double Descent
Mohammad Pezeshki
Amartya Mitra
Yoshua Bengio
Guillaume Lajoie
58
25
0
06 Dec 2021
KNAS: Green Neural Architecture Search
Jingjing Xu
Liang Zhao
Junyang Lin
Rundong Gao
Xu Sun
Hongxia Yang
17
55
0
26 Nov 2021
Model, sample, and epoch-wise descents: exact solution of gradient flow in the random feature model
A. Bodin
N. Macris
27
13
0
22 Oct 2021
Conditioning of Random Feature Matrices: Double Descent and Generalization Error
Zhijun Chen
Hayden Schaeffer
35
12
0
21 Oct 2021
Random matrices in service of ML footprint: ternary random features with no performance loss
Hafiz Tiomoko Ali
Zhenyu Liao
Romain Couillet
36
7
0
05 Oct 2021
Stochastic Training is Not Necessary for Generalization
Jonas Geiping
Micah Goldblum
Phillip E. Pope
Michael Moeller
Tom Goldstein
81
72
0
29 Sep 2021
Scaling Laws for Neural Machine Translation
Behrooz Ghorbani
Orhan Firat
Markus Freitag
Ankur Bapna
M. Krikun
Xavier Garcia
Ciprian Chelba
Colin Cherry
32
99
0
16 Sep 2021
A Farewell to the Bias-Variance Tradeoff? An Overview of the Theory of Overparameterized Machine Learning
Yehuda Dar
Vidya Muthukumar
Richard G. Baraniuk
29
71
0
06 Sep 2021
When and how epochwise double descent happens
Cory Stephenson
Tyler Lee
26
15
0
26 Aug 2021
A Random Matrix Perspective on Random Tensors
J. H. D. M. Goulart
Romain Couillet
P. Comon
11
22
0
02 Aug 2021
An Instance-Dependent Simulation Framework for Learning with Label Noise
Keren Gu
Xander Masotto
Vandana Bachani
Balaji Lakshminarayanan
Jack Nikodem
Dong Yin
NoLa
11
24
0
23 Jul 2021
Edge of chaos as a guiding principle for modern neural network training
Lin Zhang
Ling Feng
Kan Chen
C. Lai
11
9
0
20 Jul 2021
Continuous vs. Discrete Optimization of Deep Neural Networks
Omer Elkabetz
Nadav Cohen
60
44
0
14 Jul 2021
The Foes of Neural Network's Data Efficiency Among Unnecessary Input Dimensions
Vanessa D’Amario
S. Srivastava
Tomotake Sasaki
Xavier Boix
AAML
21
2
0
13 Jul 2021
Continual Learning in the Teacher-Student Setup: Impact of Task Similarity
Sebastian Lee
Sebastian Goldt
Andrew M. Saxe
CLL
24
73
0
09 Jul 2021
Mitigating deep double descent by concatenating inputs
John Chen
Qihan Wang
Anastasios Kyrillidis
BDL
8
3
0
02 Jul 2021
Saddle-to-Saddle Dynamics in Deep Linear Networks: Small Initialization Training, Symmetry, and Sparsity
Arthur Jacot
François Ged
Berfin cSimcsek
Clément Hongler
Franck Gabriel
16
52
0
30 Jun 2021
On the Sample Complexity and Metastability of Heavy-tailed Policy Search in Continuous Control
Amrit Singh Bedi
Anjaly Parayil
Junyu Zhang
Mengdi Wang
Alec Koppel
25
15
0
15 Jun 2021
An Analytical Theory of Curriculum Learning in Teacher-Student Networks
Luca Saglietti
Stefano Sarao Mannelli
Andrew M. Saxe
12
25
0
15 Jun 2021
Probing transfer learning with a model of synthetic correlated datasets
Federica Gerace
Luca Saglietti
Stefano Sarao Mannelli
Andrew M. Saxe
Lenka Zdeborová
OOD
11
30
0
09 Jun 2021
Redundant representations help generalization in wide neural networks
Diego Doimo
Aldo Glielmo
Sebastian Goldt
A. Laio
AI4CE
17
9
0
07 Jun 2021
Double Descent Optimization Pattern and Aliasing: Caveats of Noisy Labels
Florian Dubost
Erin Hong
Max Pike
Siddharth Sharma
Siyi Tang
Nandita Bhaskhar
Christopher Lee-Messer
D. Rubin
NoLa
32
0
0
03 Jun 2021
Optimization Variance: Exploring Generalization Properties of DNNs
Xiao Zhang
Dongrui Wu
Haoyi Xiong
Bo Dai
13
4
0
03 Jun 2021
Generalization Error Rates in Kernel Regression: The Crossover from the Noiseless to Noisy Regime
Hugo Cui
Bruno Loureiro
Florent Krzakala
Lenka Zdeborová
24
82
0
31 May 2021
Previous
1
2
3
4
5
6
Next