Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
1608.04414
Cited By
v1
v2
v3 (latest)
Generalization of ERM in Stochastic Convex Optimization: The Dimension Strikes Back
Neural Information Processing Systems (NeurIPS), 2016
15 August 2016
Vitaly Feldman
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Generalization of ERM in Stochastic Convex Optimization: The Dimension Strikes Back"
44 / 44 papers shown
Sample Complexity of Agnostic Multiclass Classification: Natarajan Dimension Strikes Back
Alon Cohen
Liad Erez
Steve Hanneke
Tomer Koren
Yishay Mansour
Shay Moran
Qian Zhang
113
0
0
16 Nov 2025
Flat Minima and Generalization: Insights from Stochastic Convex Optimization
Matan Schliserman
Shira Vansover-Hager
Tomer Koren
151
0
0
05 Nov 2025
Rapid Overfitting of Multi-Pass Stochastic Gradient Descent in Stochastic Convex Optimization
Shira Vansover-Hager
Tomer Koren
Roi Livni
247
1
0
13 May 2025
Characterizing the Accuracy-Communication-Privacy Trade-off in Distributed Stochastic Convex Optimization
International Conference on Artificial Intelligence and Statistics (AISTATS), 2025
Sudeep Salgia
Nikola Pavlovic
Yuejie Chi
Qing Zhao
315
0
0
06 Jan 2025
Optimal Rates for Robust Stochastic Convex Optimization
Symposium on Foundations of Responsible Computing (FRC), 2024
Changyu Gao
Andrew Lowy
Xingyu Zhou
Stephen J. Wright
578
1
0
15 Dec 2024
Information Complexity of Stochastic Convex Optimization: Applications to Generalization and Memorization
Idan Attias
Gintare Karolina Dziugaite
Mahdi Haghifam
Roi Livni
Daniel M. Roy
368
11
0
14 Feb 2024
The Sample Complexity Of ERMs In Stochastic Convex Optimization
Dan Carmon
Roi Livni
Amir Yehudayoff
295
6
0
09 Nov 2023
Initialization-Dependent Sample Complexity of Linear Predictors and Neural Networks
Neural Information Processing Systems (NeurIPS), 2023
Roey Magen
Ohad Shamir
271
2
0
25 May 2023
Select without Fear: Almost All Mini-Batch Schedules Generalize Optimally
Konstantinos E. Nikolakakis
Amin Karbasi
Dionysis Kalogerias
394
7
0
03 May 2023
Lower Generalization Bounds for GD and SGD in Smooth Stochastic Convex Optimization
Peiyuan Zhang
Jiaye Teng
J.N. Zhang
383
6
0
19 Mar 2023
Information Theoretic Lower Bounds for Information Theoretic Upper Bounds
Neural Information Processing Systems (NeurIPS), 2023
Roi Livni
325
19
0
09 Feb 2023
Generalization Bounds for Stochastic Gradient Descent via Localized
ε
\varepsilon
ε
-Covers
Neural Information Processing Systems (NeurIPS), 2022
Sejun Park
Umut Simsekli
Murat A. Erdogdu
230
10
0
19 Sep 2022
Uniform Stability for First-Order Empirical Risk Minimization
Annual Conference Computational Learning Theory (COLT), 2022
Amit Attia
Tomer Koren
242
8
0
17 Jul 2022
Beyond Lipschitz: Sharp Generalization and Excess Risk Bounds for Full-Batch GD
International Conference on Learning Representations (ICLR), 2022
Konstantinos E. Nikolakakis
Farzin Haddadpour
Amin Karbasi
Dionysios S. Kalogerias
550
23
0
26 Apr 2022
Private Convex Optimization via Exponential Mechanism
Annual Conference Computational Learning Theory (COLT), 2022
Sivakanth Gopi
Y. Lee
Daogao Liu
434
60
0
01 Mar 2022
Benign Underfitting of Stochastic Gradient Descent
Neural Information Processing Systems (NeurIPS), 2022
Tomer Koren
Roi Livni
Yishay Mansour
Uri Sherman
MLT
417
23
0
27 Feb 2022
Thinking Outside the Ball: Optimal Learning with Gradient Descent for Generalized Linear Stochastic Convex Optimization
Neural Information Processing Systems (NeurIPS), 2022
I Zaghloul Amir
Roi Livni
Nathan Srebro
320
7
0
27 Feb 2022
Black-Box Generalization: Stability of Zeroth-Order Learning
Neural Information Processing Systems (NeurIPS), 2022
Konstantinos E. Nikolakakis
Farzin Haddadpour
Dionysios S. Kalogerias
Amin Karbasi
MLT
285
2
0
14 Feb 2022
SGD: The Role of Implicit Regularization, Batch-size and Multiple-epochs
Satyen Kale
Ayush Sekhari
Karthik Sridharan
549
34
0
11 Jul 2021
Never Go Full Batch (in Stochastic Convex Optimization)
Neural Information Processing Systems (NeurIPS), 2021
I Zaghloul Amir
Y. Carmon
Tomer Koren
Roi Livni
267
16
0
29 Jun 2021
Private Non-smooth Empirical Risk Minimization and Stochastic Convex Optimization in Subquadratic Steps
Janardhan Kulkarni
Y. Lee
Daogao Liu
266
32
0
29 Mar 2021
Stability and Deviation Optimal Risk Bounds with Convergence Rate
O
(
1
/
n
)
O(1/n)
O
(
1/
n
)
Neural Information Processing Systems (NeurIPS), 2021
Yegor Klochkov
Nikita Zhivotovskiy
356
65
0
22 Mar 2021
Remember What You Want to Forget: Algorithms for Machine Unlearning
Neural Information Processing Systems (NeurIPS), 2021
Ayush Sekhari
Jayadev Acharya
Gautam Kamath
A. Suresh
FedML
MU
478
430
0
04 Mar 2021
Private Stochastic Convex Optimization: Optimal Rates in
ℓ
1
\ell_1
ℓ
1
Geometry
International Conference on Machine Learning (ICML), 2021
Hilal Asi
Vitaly Feldman
Tomer Koren
Kunal Talwar
257
104
0
02 Mar 2021
Algorithmic Instabilities of Accelerated Gradient Descent
Neural Information Processing Systems (NeurIPS), 2021
Amit Attia
Tomer Koren
239
13
0
03 Feb 2021
SGD Generalizes Better Than GD (And Regularization Doesn't Help)
Annual Conference Computational Learning Theory (COLT), 2021
I Zaghloul Amir
Tomer Koren
Roi Livni
296
52
0
01 Feb 2021
Complementary Composite Minimization, Small Gradients in General Norms, and Applications
Mathematical programming (Math. Program.), 2021
Jelena Diakonikolas
Cristóbal Guzmán
163
15
0
26 Jan 2021
Better scalability under potentially heavy-tailed feedback
Matthew J. Holland
256
1
0
14 Dec 2020
Stability of Stochastic Gradient Descent on Nonsmooth Convex Losses
Neural Information Processing Systems (NeurIPS), 2020
Raef Bassily
Vitaly Feldman
Cristóbal Guzmán
Kunal Talwar
MLT
348
224
0
12 Jun 2020
Improved scalability under heavy tails, without strong convexity
Matthew J. Holland
278
1
0
02 Jun 2020
Better scalability under potentially heavy-tailed gradients
Matthew J. Holland
309
1
0
01 Jun 2020
Private Stochastic Convex Optimization: Optimal Rates in Linear Time
Vitaly Feldman
Tomer Koren
Kunal Talwar
316
227
0
10 May 2020
Can Implicit Bias Explain Generalization? Stochastic Convex Optimization as a Case Study
Neural Information Processing Systems (NeurIPS), 2020
Assaf Dauber
M. Feder
Tomer Koren
Roi Livni
298
26
0
13 Mar 2020
Reasoning About Generalization via Conditional Mutual Information
Annual Conference Computational Learning Theory (COLT), 2020
Thomas Steinke
Lydia Zakynthinou
527
195
0
24 Jan 2020
Private Stochastic Convex Optimization with Optimal Rates
Neural Information Processing Systems (NeurIPS), 2019
Raef Bassily
Vitaly Feldman
Kunal Talwar
Abhradeep Thakurta
359
267
0
27 Aug 2019
Does Learning Require Memorization? A Short Tale about a Long Tail
Symposium on the Theory of Computing (STOC), 2019
Vitaly Feldman
TDI
788
615
0
12 Jun 2019
High probability generalization bounds for uniformly stable algorithms with nearly optimal rate
Annual Conference Computational Learning Theory (COLT), 2019
Vitaly Feldman
J. Vondrák
315
173
0
27 Feb 2019
Stochastic Approximation of Smooth and Strongly Convex Functions: Beyond the
O
(
1
/
T
)
O(1/T)
O
(
1/
T
)
Convergence Rate
Lijun Zhang
Zhi Zhou
278
32
0
27 Jan 2019
Generalization Bounds for Uniformly Stable Algorithms
Vitaly Feldman
J. Vondrák
230
98
0
24 Dec 2018
Robust descent using smoothed multiplicative noise
Matthew J. Holland
OOD
191
28
0
15 Oct 2018
Fast Rates of ERM and Stochastic Approximation: Adaptive to Error Bound Conditions
Mingrui Liu
Xiaoxuan Zhang
Lijun Zhang
Rong Jin
Tianbao Yang
261
28
0
11 May 2018
A Simple Analysis for Exp-concave Empirical Minimization with Arbitrary Convex Regularizer
Tianbao Yang
Zhe Li
Lijun Zhang
166
7
0
09 Sep 2017
Efficient learning with robust gradient descent
Machine-mediated learning (ML), 2017
Matthew J. Holland
K. Ikeda
OOD
416
28
0
01 Jun 2017
Empirical Risk Minimization for Stochastic Convex Optimization:
O
(
1
/
n
)
O(1/n)
O
(
1/
n
)
- and
O
(
1
/
n
2
)
O(1/n^2)
O
(
1/
n
2
)
-type of Risk Bounds
Annual Conference Computational Learning Theory (COLT), 2017
Lijun Zhang
Tianbao Yang
Rong Jin
256
52
0
07 Feb 2017
1
Page 1 of 1