ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2008.01036
  4. Cited By
Multiple Descent: Design Your Own Generalization Curve
v1v2v3v4v5v6v7 (latest)

Multiple Descent: Design Your Own Generalization Curve

3 August 2020
Lin Chen
Yifei Min
M. Belkin
Amin Karbasi
    DRL
ArXiv (abs)PDFHTML

Papers citing "Multiple Descent: Design Your Own Generalization Curve"

45 / 45 papers shown
A dynamic view of some anomalous phenomena in SGD
A dynamic view of some anomalous phenomena in SGD
Vivek Shripad Borkar
438
0
0
03 May 2025
Investigating the Impact of Model Complexity in Large Language Models
Investigating the Impact of Model Complexity in Large Language Models
Jing Luo
Huiyuan Wang
Weiran Huang
262
0
0
01 Oct 2024
Understanding the Double Descent Phenomenon in Deep Learning
Understanding the Double Descent Phenomenon in Deep Learning
Marc Lafon
Alexandre Thomas
428
4
0
15 Mar 2024
A PAC-Bayesian Perspective on the Interpolating Information Criterion
A PAC-Bayesian Perspective on the Interpolating Information Criterion
Liam Hodgkinson
Christopher van der Heide
Roberto Salomone
Fred Roosta
Michael W. Mahoney
334
3
0
13 Nov 2023
Transgressing the boundaries: towards a rigorous understanding of deep
  learning and its (non-)robustness
Transgressing the boundaries: towards a rigorous understanding of deep learning and its (non-)robustness
C. Hartmann
Lorenz Richter
AAML
248
2
0
05 Jul 2023
Predicting Grokking Long Before it Happens: A look into the loss
  landscape of models which grok
Predicting Grokking Long Before it Happens: A look into the loss landscape of models which grok
Pascal Junior Tikeng Notsawo
Hattie Zhou
Mohammad Pezeshki
Irina Rish
G. Dumas
343
30
0
23 Jun 2023
Dropout Drops Double Descent
Dropout Drops Double DescentJapanese Journal of Statistics and Data Science (JSDS), 2023
Tianbao Yang
J. Suzuki
342
1
0
25 May 2023
Least Squares Regression Can Exhibit Under-Parameterized Double Descent
Least Squares Regression Can Exhibit Under-Parameterized Double DescentNeural Information Processing Systems (NeurIPS), 2023
Xinyue Li
Rishi Sonthalia
437
5
0
24 May 2023
Double Descent Demystified: Identifying, Interpreting & Ablating the
  Sources of a Deep Learning Puzzle
Double Descent Demystified: Identifying, Interpreting & Ablating the Sources of a Deep Learning Puzzle
Rylan Schaeffer
Mikail Khona
Zachary Robertson
Akhilan Boopathy
Kateryna Pistunova
J. Rocks
Ila Rani Fiete
Oluwasanmi Koyejo
353
48
0
24 Mar 2023
Extrapolated cross-validation for randomized ensembles
Extrapolated cross-validation for randomized ensemblesJournal of Computational And Graphical Statistics (JCGS), 2023
Jin-Hong Du
Pratik V. Patil
Kathryn Roeder
Arun K. Kuchibhotla
524
9
0
27 Feb 2023
Finding Regularized Competitive Equilibria of Heterogeneous Agent
  Macroeconomic Models with Reinforcement Learning
Finding Regularized Competitive Equilibria of Heterogeneous Agent Macroeconomic Models with Reinforcement LearningInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2023
Ruitu Xu
Yifei Min
Tianhao Wang
Zhaoran Wang
Michael I. Jordan
Zhuoran Yang
253
8
0
24 Feb 2023
Deep networks for system identification: a Survey
Deep networks for system identification: a Survey
G. Pillonetto
Aleksandr Aravkin
Daniel Gedon
L. Ljung
Antônio H. Ribeiro
Thomas B. Schon
OOD
372
104
0
30 Jan 2023
Homophily modulates double descent generalization in graph convolution
  networks
Homophily modulates double descent generalization in graph convolution networksProceedings of the National Academy of Sciences of the United States of America (PNAS), 2022
Chengzhi Shi
Liming Pan
Hong Hu
Ivan Dokmanić
535
13
0
26 Dec 2022
Gradient flow in the gaussian covariate model: exact solution of
  learning curves and multiple descent structures
Gradient flow in the gaussian covariate model: exact solution of learning curves and multiple descent structures
Antione Bodin
N. Macris
312
5
0
13 Dec 2022
Regularization Trade-offs with Fake Features
Regularization Trade-offs with Fake FeaturesEuropean Signal Processing Conference (EUSIPCO), 2022
Martin Hellkvist
Ayça Özçelikkale
Anders Ahlén
388
0
0
01 Dec 2022
A Non-Asymptotic Moreau Envelope Theory for High-Dimensional Generalized
  Linear Models
A Non-Asymptotic Moreau Envelope Theory for High-Dimensional Generalized Linear ModelsNeural Information Processing Systems (NeurIPS), 2022
Lijia Zhou
Frederic Koehler
Pragya Sur
Danica J. Sutherland
Nathan Srebro
394
12
0
21 Oct 2022
Second-order regression models exhibit progressive sharpening to the
  edge of stability
Second-order regression models exhibit progressive sharpening to the edge of stabilityInternational Conference on Machine Learning (ICML), 2022
Atish Agarwala
Fabian Pedregosa
Jeffrey Pennington
323
34
0
10 Oct 2022
Omnigrok: Grokking Beyond Algorithmic Data
Omnigrok: Grokking Beyond Algorithmic DataInternational Conference on Learning Representations (ICLR), 2022
Ziming Liu
Eric J. Michaud
Max Tegmark
441
126
0
03 Oct 2022
Multiple Descent in the Multiple Random Feature Model
Multiple Descent in the Multiple Random Feature ModelJournal of machine learning research (JMLR), 2022
Xuran Meng
Jianfeng Yao
Yuan Cao
302
10
0
21 Aug 2022
Information bottleneck theory of high-dimensional regression: relevancy,
  efficiency and optimality
Information bottleneck theory of high-dimensional regression: relevancy, efficiency and optimalityNeural Information Processing Systems (NeurIPS), 2022
Wave Ngampruetikorn
David J. Schwab
216
10
0
08 Aug 2022
The BUTTER Zone: An Empirical Study of Training Dynamics in Fully
  Connected Neural Networks
The BUTTER Zone: An Empirical Study of Training Dynamics in Fully Connected Neural Networks
Charles Edison Tripp
J. Perr-Sauer
L. Hayne
M. Lunacek
Jamil Gafur
AI4CE
384
2
0
25 Jul 2022
Improving Students' Academic Performance with AI and Semantic
  Technologies
Improving Students' Academic Performance with AI and Semantic Technologies
Yi-Hua Cheng
169
5
0
02 May 2022
Cascaded Gaps: Towards Gap-Dependent Regret for Risk-Sensitive
  Reinforcement Learning
Cascaded Gaps: Towards Gap-Dependent Regret for Risk-Sensitive Reinforcement Learning
Yingjie Fei
Ruitu Xu
196
5
0
07 Mar 2022
Benefit of Interpolation in Nearest Neighbor Algorithms
Benefit of Interpolation in Nearest Neighbor AlgorithmsSIAM Journal on Mathematics of Data Science (SIMODS), 2019
Yue Xing
Qifan Song
Guang Cheng
347
44
0
23 Feb 2022
On Optimal Early Stopping: Over-informative versus Under-informative
  Parametrization
On Optimal Early Stopping: Over-informative versus Under-informative Parametrization
Ruoqi Shen
Liyao (Mars) Gao
Yi-An Ma
351
16
0
20 Feb 2022
Fluctuations, Bias, Variance & Ensemble of Learners: Exact Asymptotics
  for Convex Losses in High-Dimension
Fluctuations, Bias, Variance & Ensemble of Learners: Exact Asymptotics for Convex Losses in High-DimensionInternational Conference on Machine Learning (ICML), 2022
Bruno Loureiro
Cédric Gerbelot
Maria Refinetti
G. Sicuro
Florent Krzakala
316
28
0
31 Jan 2022
With Greater Distance Comes Worse Performance: On the Perspective of
  Layer Utilization and Model Generalization
With Greater Distance Comes Worse Performance: On the Perspective of Layer Utilization and Model Generalization
James Wang
Cheng Yang
105
0
0
28 Jan 2022
Multi-scale Feature Learning Dynamics: Insights for Double Descent
Multi-scale Feature Learning Dynamics: Insights for Double DescentInternational Conference on Machine Learning (ICML), 2021
Mohammad Pezeshki
Amartya Mitra
Yoshua Bengio
Guillaume Lajoie
271
32
0
06 Dec 2021
Exponential Bellman Equation and Improved Regret Bounds for
  Risk-Sensitive Reinforcement Learning
Exponential Bellman Equation and Improved Regret Bounds for Risk-Sensitive Reinforcement Learning
Yingjie Fei
Zhuoran Yang
Yudong Chen
Zhaoran Wang
347
62
0
06 Nov 2021
On the Double Descent of Random Features Models Trained with SGD
On the Double Descent of Random Features Models Trained with SGD
Fanghui Liu
Johan A. K. Suykens
Volkan Cevher
MLT
552
11
0
13 Oct 2021
Taxonomizing local versus global structure in neural network loss
  landscapes
Taxonomizing local versus global structure in neural network loss landscapesNeural Information Processing Systems (NeurIPS), 2021
Yaoqing Yang
Liam Hodgkinson
Ryan Theisen
Joe Zou
Joseph E. Gonzalez
Kannan Ramchandran
Michael W. Mahoney
408
46
0
23 Jul 2021
Mitigating deep double descent by concatenating inputs
Mitigating deep double descent by concatenating inputs
John Chen
Qihan Wang
Anastasios Kyrillidis
BDL
230
3
0
02 Jul 2021
Towards an Understanding of Benign Overfitting in Neural Networks
Towards an Understanding of Benign Overfitting in Neural Networks
Zhu Li
Zhi Zhou
Arthur Gretton
MLT
314
35
0
06 Jun 2021
Out-of-Distribution Generalization in Kernel Regression
Out-of-Distribution Generalization in Kernel RegressionNeural Information Processing Systems (NeurIPS), 2021
Abdulkadir Canatar
Blake Bordelon
Cengiz Pehlevan
OODDOOD
253
21
0
04 Jun 2021
The Shape of Learning Curves: a Review
The Shape of Learning Curves: a ReviewIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2021
T. Viering
Marco Loog
395
198
0
19 Mar 2021
On the interplay between data structure and loss function in
  classification problems
On the interplay between data structure and loss function in classification problemsNeural Information Processing Systems (NeurIPS), 2021
Stéphane dÁscoli
Marylou Gabrié
Levent Sagun
Giulio Biroli
326
17
0
09 Mar 2021
Risk-Monotonicity in Statistical Learning
Risk-Monotonicity in Statistical LearningNeural Information Processing Systems (NeurIPS), 2020
Zakaria Mhammedi
656
9
0
28 Nov 2020
What causes the test error? Going beyond bias-variance via ANOVA
What causes the test error? Going beyond bias-variance via ANOVAJournal of machine learning research (JMLR), 2020
Licong Lin
Guang Cheng
324
36
0
11 Oct 2020
On the Universality of the Double Descent Peak in Ridgeless Regression
On the Universality of the Double Descent Peak in Ridgeless RegressionInternational Conference on Learning Representations (ICLR), 2020
David Holzmüller
605
15
0
05 Oct 2020
Spectral Bias and Task-Model Alignment Explain Generalization in Kernel
  Regression and Infinitely Wide Neural Networks
Spectral Bias and Task-Model Alignment Explain Generalization in Kernel Regression and Infinitely Wide Neural NetworksNature Communications (Nat Commun), 2020
Abdulkadir Canatar
Blake Bordelon
Cengiz Pehlevan
637
234
0
23 Jun 2020
Asymptotics of Ridge (less) Regression under General Source Condition
Asymptotics of Ridge (less) Regression under General Source ConditionInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2020
Dominic Richards
Jaouad Mourtada
Lorenzo Rosasco
393
85
0
11 Jun 2020
Triple descent and the two kinds of overfitting: Where & why do they
  appear?
Triple descent and the two kinds of overfitting: Where & why do they appear?
Stéphane dÁscoli
Levent Sagun
Giulio Biroli
341
84
0
05 Jun 2020
Random Features for Kernel Approximation: A Survey on Algorithms,
  Theory, and Beyond
Random Features for Kernel Approximation: A Survey on Algorithms, Theory, and Beyond
Fanghui Liu
Xiaolin Huang
Yudong Chen
Johan A. K. Suykens
BDL
565
200
0
23 Apr 2020
The Curious Case of Adversarially Robust Models: More Data Can Help,
  Double Descend, or Hurt Generalization
The Curious Case of Adversarially Robust Models: More Data Can Help, Double Descend, or Hurt GeneralizationConference on Uncertainty in Artificial Intelligence (UAI), 2020
Yifei Min
Lin Chen
Amin Karbasi
AAML
349
72
0
25 Feb 2020
Empirical Risk Minimization in the Interpolating Regime with Application
  to Neural Network Learning
Empirical Risk Minimization in the Interpolating Regime with Application to Neural Network LearningMachine-mediated learning (ML), 2019
Nicole Mücke
Ingo Steinwart
AI4CE
280
3
0
25 May 2019
1
Page 1 of 1