Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2003.06152
Cited By
v1
v2
v3 (latest)
Can Implicit Bias Explain Generalization? Stochastic Convex Optimization as a Case Study
Neural Information Processing Systems (NeurIPS), 2020
13 March 2020
Assaf Dauber
M. Feder
Tomer Koren
Roi Livni
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Can Implicit Bias Explain Generalization? Stochastic Convex Optimization as a Case Study"
36 / 36 papers shown
Make Optimization Once and for All with Fine-grained Guidance
Mingjia Shi
Ruihan Lin
Xuxi Chen
Yuhao Zhou
Zezhen Ding
...
Tong Wang
Xiaojiang Peng
Zhangyang Wang
Jing Zhang
Tianlong Chen
348
1
0
14 Mar 2025
Improving Implicit Regularization of SGD with Preconditioning for Least Square Problems
Junwei Su
Difan Zou
Chuan Wu
476
0
0
13 Mar 2024
Information Complexity of Stochastic Convex Optimization: Applications to Generalization and Memorization
Idan Attias
Gintare Karolina Dziugaite
Mahdi Haghifam
Roi Livni
Daniel M. Roy
383
13
0
14 Feb 2024
The Sample Complexity Of ERMs In Stochastic Convex Optimization
Dan Carmon
Roi Livni
Amir Yehudayoff
304
6
0
09 Nov 2023
Towards understanding deep learning with the natural clustering prior
Simon Carbonnelle
246
0
0
15 Mar 2022
Benign Underfitting of Stochastic Gradient Descent
Neural Information Processing Systems (NeurIPS), 2022
Tomer Koren
Roi Livni
Yishay Mansour
Uri Sherman
MLT
422
23
0
27 Feb 2022
Thinking Outside the Ball: Optimal Learning with Gradient Descent for Generalized Linear Stochastic Convex Optimization
Neural Information Processing Systems (NeurIPS), 2022
I Zaghloul Amir
Roi Livni
Nathan Srebro
327
7
0
27 Feb 2022
Limitation of Characterizing Implicit Regularization by Data-independent Functions
Leyang Zhang
Z. Xu
Yaoyu Zhang
Yaoyu Zhang
241
0
0
28 Jan 2022
Stochastic Training is Not Necessary for Generalization
Jonas Geiping
Micah Goldblum
Phillip E. Pope
Michael Moeller
Tom Goldstein
534
82
0
29 Sep 2021
The Benefits of Implicit Regularization from SGD in Least Squares Problems
Neural Information Processing Systems (NeurIPS), 2021
Difan Zou
Jingfeng Wu
Vladimir Braverman
Quanquan Gu
Dean Phillips Foster
Sham Kakade
200
37
0
10 Aug 2021
SGD: The Role of Implicit Regularization, Batch-size and Multiple-epochs
Satyen Kale
Ayush Sekhari
Karthik Sridharan
565
34
0
11 Jul 2021
Never Go Full Batch (in Stochastic Convex Optimization)
Neural Information Processing Systems (NeurIPS), 2021
I Zaghloul Amir
Y. Carmon
Tomer Koren
Roi Livni
271
16
0
29 Jun 2021
SGD Generalizes Better Than GD (And Regularization Doesn't Help)
Annual Conference Computational Learning Theory (COLT), 2021
I Zaghloul Amir
Tomer Koren
Roi Livni
297
52
0
01 Feb 2021
Implicit Regularization in ReLU Networks with the Square Loss
Annual Conference Computational Learning Theory (COLT), 2020
Gal Vardi
Ohad Shamir
298
53
0
09 Dec 2020
On Convergence and Generalization of Dropout Training
Neural Information Processing Systems (NeurIPS), 2020
Poorya Mianjy
R. Arora
318
33
0
23 Oct 2020
Is SGD a Bayesian sampler? Well, almost
Chris Mingard
Guillermo Valle Pérez
Joar Skalse
A. Louis
BDL
485
69
0
26 Jun 2020
A Limitation of the PAC-Bayes Framework
Neural Information Processing Systems (NeurIPS), 2020
Roi Livni
Shay Moran
445
26
0
24 Jun 2020
Implicit Regularization in Deep Learning May Not Be Explainable by Norms
Noam Razin
Nadav Cohen
389
169
0
13 May 2020
Implicit Regularization in Deep Matrix Factorization
Neural Information Processing Systems (NeurIPS), 2019
Sanjeev Arora
Nadav Cohen
Wei Hu
Yuping Luo
AI4CE
555
585
0
31 May 2019
Uniform convergence may be unable to explain generalization in deep learning
Neural Information Processing Systems (NeurIPS), 2019
Vaishnavh Nagarajan
J. Zico Kolter
MoMe
AI4CE
610
351
0
13 Feb 2019
A Continuous-Time View of Early Stopping for Least Squares
Alnur Ali
J. Zico Kolter
Robert Tibshirani
356
107
0
23 Oct 2018
Implicit Bias of Gradient Descent on Linear Convolutional Networks
Suriya Gunasekar
Jason D. Lee
Daniel Soudry
Nathan Srebro
MDE
573
456
0
01 Jun 2018
Convergence of Gradient Descent on Separable Data
Mor Shpigel Nacson
Jason D. Lee
Suriya Gunasekar
Pedro H. P. Savarese
Nathan Srebro
Daniel Soudry
523
184
0
05 Mar 2018
Characterizing Implicit Bias in Terms of Optimization Geometry
Suriya Gunasekar
Jason D. Lee
Daniel Soudry
Nathan Srebro
AI4CE
603
457
0
22 Feb 2018
The Implicit Bias of Gradient Descent on Separable Data
Journal of machine learning research (JMLR), 2017
Daniel Soudry
Elad Hoffer
Mor Shpigel Nacson
Suriya Gunasekar
Nathan Srebro
1.2K
1,077
0
27 Oct 2017
Implicit Regularization in Deep Learning
Behnam Neyshabur
486
164
0
06 Sep 2017
Early stopping for kernel boosting algorithms: A general analysis with localized complexities
Yuting Wei
Fanny Yang
Martin J. Wainwright
232
86
0
05 Jul 2017
Implicit Regularization in Matrix Factorization
Suriya Gunasekar
Blake E. Woodworth
Srinadh Bhojanapalli
Behnam Neyshabur
Nathan Srebro
447
555
0
25 May 2017
Geometry of Optimization and Implicit Regularization in Deep Learning
Behnam Neyshabur
Ryota Tomioka
Ruslan Salakhutdinov
Nathan Srebro
AI4CE
248
140
0
08 May 2017
Understanding deep learning requires rethinking generalization
Chiyuan Zhang
Samy Bengio
Moritz Hardt
Benjamin Recht
Oriol Vinyals
HAI
955
5,020
0
10 Nov 2016
Generalization of ERM in Stochastic Convex Optimization: The Dimension Strikes Back
Neural Information Processing Systems (NeurIPS), 2016
Vitaly Feldman
373
75
0
15 Aug 2016
Generalization Properties and Implicit Regularization for Multiple Passes SGM
Junhong Lin
Raffaello Camoriano
Lorenzo Rosasco
203
71
0
26 May 2016
Train faster, generalize better: Stability of stochastic gradient descent
Moritz Hardt
Benjamin Recht
Y. Singer
560
1,390
0
03 Sep 2015
In Search of the Real Inductive Bias: On the Role of Implicit Regularization in Deep Learning
International Conference on Learning Representations (ICLR), 2014
Behnam Neyshabur
Ryota Tomioka
Nathan Srebro
AI4CE
725
714
0
20 Dec 2014
Early stopping and non-parametric regression: An optimal data-dependent stopping rule
Allerton Conference on Communication, Control, and Computing (Allerton), 2011
Garvesh Raskutti
Martin J. Wainwright
Bin Yu
339
327
0
15 Jun 2013
Stochastic Gradient Descent for Non-smooth Optimization: Convergence Results and Optimal Averaging Schemes
Ohad Shamir
Tong Zhang
555
619
0
08 Dec 2012
1
Page 1 of 1