ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1810.13395
  4. Cited By
Accelerating SGD with momentum for over-parameterized learning
v1v2v3v4v5 (latest)

Accelerating SGD with momentum for over-parameterized learning

31 October 2018
Chaoyue Liu
M. Belkin
    ODL
ArXiv (abs)PDFHTML

Papers citing "Accelerating SGD with momentum for over-parameterized learning"

12 / 12 papers shown
Nesterov acceleration in benignly non-convex landscapes
Nesterov acceleration in benignly non-convex landscapesInternational Conference on Learning Representations (ICLR), 2024
Kanan Gupta
Stephan Wojtowytsch
306
4
0
10 Oct 2024
Accelerated stochastic approximation with state-dependent noise
Accelerated stochastic approximation with state-dependent noiseMathematical programming (Math. Program.), 2023
Sasila Ilandarideva
A. Juditsky
Guanghui Lan
Tianjiao Li
387
12
0
04 Jul 2023
Adaptive Step-Size Methods for Compressed SGD
Adaptive Step-Size Methods for Compressed SGDIEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2022
Adarsh M. Subramaniam
A. Magesh
Venugopal V. Veeravalli
146
1
0
20 Jul 2022
Nesterov Accelerated Shuffling Gradient Method for Convex Optimization
Nesterov Accelerated Shuffling Gradient Method for Convex OptimizationInternational Conference on Machine Learning (ICML), 2022
Trang H. Tran
K. Scheinberg
Lam M. Nguyen
364
14
0
07 Feb 2022
An Even More Optimal Stochastic Optimization Algorithm: Minibatching and
  Interpolation Learning
An Even More Optimal Stochastic Optimization Algorithm: Minibatching and Interpolation LearningNeural Information Processing Systems (NeurIPS), 2021
Blake E. Woodworth
Nathan Srebro
240
22
0
04 Jun 2021
A Unified Analysis of First-Order Methods for Smooth Games via Integral
  Quadratic Constraints
A Unified Analysis of First-Order Methods for Smooth Games via Integral Quadratic ConstraintsJournal of machine learning research (JMLR), 2020
Guodong Zhang
Xuchao Bao
Laurent Lessard
Roger C. Grosse
365
26
0
23 Sep 2020
On the Generalization Benefit of Noise in Stochastic Gradient Descent
On the Generalization Benefit of Noise in Stochastic Gradient Descent
Samuel L. Smith
Erich Elsen
Soham De
MLT
217
116
0
26 Jun 2020
DEED: A General Quantization Scheme for Communication Efficiency in Bits
DEED: A General Quantization Scheme for Communication Efficiency in Bits
Tian-Chun Ye
Peijun Xiao
Tian Ding
FedMLMQ
93
2
0
19 Jun 2020
Optimization for deep learning: theory and algorithms
Optimization for deep learning: theory and algorithms
Tian Ding
ODL
343
179
0
19 Dec 2019
Student Specialization in Deep ReLU Networks With Finite Width and Input
  Dimension
Student Specialization in Deep ReLU Networks With Finite Width and Input Dimension
Yuandong Tian
MLT
218
8
0
30 Sep 2019
Painless Stochastic Gradient: Interpolation, Line-Search, and
  Convergence Rates
Painless Stochastic Gradient: Interpolation, Line-Search, and Convergence RatesNeural Information Processing Systems (NeurIPS), 2019
Sharan Vaswani
Aaron Mishkin
I. Laradji
Mark Schmidt
Gauthier Gidel
Damien Scieur
ODL
441
231
0
24 May 2019
Fast and Faster Convergence of SGD for Over-Parameterized Models and an
  Accelerated Perceptron
Fast and Faster Convergence of SGD for Over-Parameterized Models and an Accelerated Perceptron
Sharan Vaswani
Francis R. Bach
Mark Schmidt
398
320
0
16 Oct 2018
1
Page 1 of 1