ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2003.06152
  4. Cited By
Can Implicit Bias Explain Generalization? Stochastic Convex Optimization
  as a Case Study
v1v2v3 (latest)

Can Implicit Bias Explain Generalization? Stochastic Convex Optimization as a Case Study

Neural Information Processing Systems (NeurIPS), 2020
13 March 2020
Assaf Dauber
M. Feder
Tomer Koren
Roi Livni
ArXiv (abs)PDFHTML

Papers citing "Can Implicit Bias Explain Generalization? Stochastic Convex Optimization as a Case Study"

36 / 36 papers shown
Make Optimization Once and for All with Fine-grained Guidance
Make Optimization Once and for All with Fine-grained Guidance
Mingjia Shi
Ruihan Lin
Xuxi Chen
Yuhao Zhou
Zezhen Ding
...
Tong Wang
Xiaojiang Peng
Zhangyang Wang
Jing Zhang
Tianlong Chen
348
1
0
14 Mar 2025
Improving Implicit Regularization of SGD with Preconditioning for Least
  Square Problems
Improving Implicit Regularization of SGD with Preconditioning for Least Square Problems
Junwei Su
Difan Zou
Chuan Wu
476
0
0
13 Mar 2024
Information Complexity of Stochastic Convex Optimization: Applications
  to Generalization and Memorization
Information Complexity of Stochastic Convex Optimization: Applications to Generalization and Memorization
Idan Attias
Gintare Karolina Dziugaite
Mahdi Haghifam
Roi Livni
Daniel M. Roy
383
13
0
14 Feb 2024
The Sample Complexity Of ERMs In Stochastic Convex Optimization
The Sample Complexity Of ERMs In Stochastic Convex Optimization
Dan Carmon
Roi Livni
Amir Yehudayoff
304
6
0
09 Nov 2023
Towards understanding deep learning with the natural clustering prior
Towards understanding deep learning with the natural clustering prior
Simon Carbonnelle
246
0
0
15 Mar 2022
Benign Underfitting of Stochastic Gradient Descent
Benign Underfitting of Stochastic Gradient DescentNeural Information Processing Systems (NeurIPS), 2022
Tomer Koren
Roi Livni
Yishay Mansour
Uri Sherman
MLT
422
23
0
27 Feb 2022
Thinking Outside the Ball: Optimal Learning with Gradient Descent for
  Generalized Linear Stochastic Convex Optimization
Thinking Outside the Ball: Optimal Learning with Gradient Descent for Generalized Linear Stochastic Convex OptimizationNeural Information Processing Systems (NeurIPS), 2022
I Zaghloul Amir
Roi Livni
Nathan Srebro
327
7
0
27 Feb 2022
Limitation of Characterizing Implicit Regularization by Data-independent
  Functions
Limitation of Characterizing Implicit Regularization by Data-independent Functions
Leyang Zhang
Z. Xu
Yaoyu Zhang
Yaoyu Zhang
241
0
0
28 Jan 2022
Stochastic Training is Not Necessary for Generalization
Stochastic Training is Not Necessary for Generalization
Jonas Geiping
Micah Goldblum
Phillip E. Pope
Michael Moeller
Tom Goldstein
534
82
0
29 Sep 2021
The Benefits of Implicit Regularization from SGD in Least Squares
  Problems
The Benefits of Implicit Regularization from SGD in Least Squares ProblemsNeural Information Processing Systems (NeurIPS), 2021
Difan Zou
Jingfeng Wu
Vladimir Braverman
Quanquan Gu
Dean Phillips Foster
Sham Kakade
200
37
0
10 Aug 2021
SGD: The Role of Implicit Regularization, Batch-size and Multiple-epochs
SGD: The Role of Implicit Regularization, Batch-size and Multiple-epochs
Satyen Kale
Ayush Sekhari
Karthik Sridharan
565
34
0
11 Jul 2021
Never Go Full Batch (in Stochastic Convex Optimization)
Never Go Full Batch (in Stochastic Convex Optimization)Neural Information Processing Systems (NeurIPS), 2021
I Zaghloul Amir
Y. Carmon
Tomer Koren
Roi Livni
271
16
0
29 Jun 2021
SGD Generalizes Better Than GD (And Regularization Doesn't Help)
SGD Generalizes Better Than GD (And Regularization Doesn't Help)Annual Conference Computational Learning Theory (COLT), 2021
I Zaghloul Amir
Tomer Koren
Roi Livni
297
52
0
01 Feb 2021
Implicit Regularization in ReLU Networks with the Square Loss
Implicit Regularization in ReLU Networks with the Square LossAnnual Conference Computational Learning Theory (COLT), 2020
Gal Vardi
Ohad Shamir
298
53
0
09 Dec 2020
On Convergence and Generalization of Dropout Training
On Convergence and Generalization of Dropout TrainingNeural Information Processing Systems (NeurIPS), 2020
Poorya Mianjy
R. Arora
318
33
0
23 Oct 2020
Is SGD a Bayesian sampler? Well, almost
Is SGD a Bayesian sampler? Well, almost
Chris Mingard
Guillermo Valle Pérez
Joar Skalse
A. Louis
BDL
485
69
0
26 Jun 2020
A Limitation of the PAC-Bayes Framework
A Limitation of the PAC-Bayes FrameworkNeural Information Processing Systems (NeurIPS), 2020
Roi Livni
Shay Moran
445
26
0
24 Jun 2020
Implicit Regularization in Deep Learning May Not Be Explainable by Norms
Implicit Regularization in Deep Learning May Not Be Explainable by Norms
Noam Razin
Nadav Cohen
389
169
0
13 May 2020
Implicit Regularization in Deep Matrix Factorization
Implicit Regularization in Deep Matrix FactorizationNeural Information Processing Systems (NeurIPS), 2019
Sanjeev Arora
Nadav Cohen
Wei Hu
Yuping Luo
AI4CE
555
585
0
31 May 2019
Uniform convergence may be unable to explain generalization in deep
  learning
Uniform convergence may be unable to explain generalization in deep learningNeural Information Processing Systems (NeurIPS), 2019
Vaishnavh Nagarajan
J. Zico Kolter
MoMeAI4CE
610
351
0
13 Feb 2019
A Continuous-Time View of Early Stopping for Least Squares
A Continuous-Time View of Early Stopping for Least Squares
Alnur Ali
J. Zico Kolter
Robert Tibshirani
356
107
0
23 Oct 2018
Implicit Bias of Gradient Descent on Linear Convolutional Networks
Implicit Bias of Gradient Descent on Linear Convolutional Networks
Suriya Gunasekar
Jason D. Lee
Daniel Soudry
Nathan Srebro
MDE
573
456
0
01 Jun 2018
Convergence of Gradient Descent on Separable Data
Convergence of Gradient Descent on Separable Data
Mor Shpigel Nacson
Jason D. Lee
Suriya Gunasekar
Pedro H. P. Savarese
Nathan Srebro
Daniel Soudry
523
184
0
05 Mar 2018
Characterizing Implicit Bias in Terms of Optimization Geometry
Characterizing Implicit Bias in Terms of Optimization Geometry
Suriya Gunasekar
Jason D. Lee
Daniel Soudry
Nathan Srebro
AI4CE
603
457
0
22 Feb 2018
The Implicit Bias of Gradient Descent on Separable Data
The Implicit Bias of Gradient Descent on Separable DataJournal of machine learning research (JMLR), 2017
Daniel Soudry
Elad Hoffer
Mor Shpigel Nacson
Suriya Gunasekar
Nathan Srebro
1.2K
1,077
0
27 Oct 2017
Implicit Regularization in Deep Learning
Implicit Regularization in Deep Learning
Behnam Neyshabur
486
164
0
06 Sep 2017
Early stopping for kernel boosting algorithms: A general analysis with
  localized complexities
Early stopping for kernel boosting algorithms: A general analysis with localized complexities
Yuting Wei
Fanny Yang
Martin J. Wainwright
232
86
0
05 Jul 2017
Implicit Regularization in Matrix Factorization
Implicit Regularization in Matrix Factorization
Suriya Gunasekar
Blake E. Woodworth
Srinadh Bhojanapalli
Behnam Neyshabur
Nathan Srebro
447
555
0
25 May 2017
Geometry of Optimization and Implicit Regularization in Deep Learning
Geometry of Optimization and Implicit Regularization in Deep Learning
Behnam Neyshabur
Ryota Tomioka
Ruslan Salakhutdinov
Nathan Srebro
AI4CE
248
140
0
08 May 2017
Understanding deep learning requires rethinking generalization
Understanding deep learning requires rethinking generalization
Chiyuan Zhang
Samy Bengio
Moritz Hardt
Benjamin Recht
Oriol Vinyals
HAI
955
5,020
0
10 Nov 2016
Generalization of ERM in Stochastic Convex Optimization: The Dimension
  Strikes Back
Generalization of ERM in Stochastic Convex Optimization: The Dimension Strikes BackNeural Information Processing Systems (NeurIPS), 2016
Vitaly Feldman
373
75
0
15 Aug 2016
Generalization Properties and Implicit Regularization for Multiple
  Passes SGM
Generalization Properties and Implicit Regularization for Multiple Passes SGM
Junhong Lin
Raffaello Camoriano
Lorenzo Rosasco
203
71
0
26 May 2016
Train faster, generalize better: Stability of stochastic gradient
  descent
Train faster, generalize better: Stability of stochastic gradient descent
Moritz Hardt
Benjamin Recht
Y. Singer
560
1,390
0
03 Sep 2015
In Search of the Real Inductive Bias: On the Role of Implicit
  Regularization in Deep Learning
In Search of the Real Inductive Bias: On the Role of Implicit Regularization in Deep LearningInternational Conference on Learning Representations (ICLR), 2014
Behnam Neyshabur
Ryota Tomioka
Nathan Srebro
AI4CE
725
714
0
20 Dec 2014
Early stopping and non-parametric regression: An optimal data-dependent
  stopping rule
Early stopping and non-parametric regression: An optimal data-dependent stopping ruleAllerton Conference on Communication, Control, and Computing (Allerton), 2011
Garvesh Raskutti
Martin J. Wainwright
Bin Yu
339
327
0
15 Jun 2013
Stochastic Gradient Descent for Non-smooth Optimization: Convergence
  Results and Optimal Averaging Schemes
Stochastic Gradient Descent for Non-smooth Optimization: Convergence Results and Optimal Averaging Schemes
Ohad Shamir
Tong Zhang
555
619
0
08 Dec 2012
1
Page 1 of 1