Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2110.03891
Cited By
Does Momentum Change the Implicit Regularization on Separable Data?
8 October 2021
Bohan Wang
Qi Meng
Huishuai Zhang
Ruoyu Sun
Wei-Neng Chen
Zhirui Ma
Tie-Yan Liu
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Does Momentum Change the Implicit Regularization on Separable Data?"
5 / 5 papers shown
Title
Gradient Descent Robustly Learns the Intrinsic Dimension of Data in Training Convolutional Neural Networks
Chenyang Zhang
Peifeng Gao
Difan Zou
Yuan Cao
OOD
MLT
54
0
0
11 Apr 2025
A Simple Convergence Proof of Adam and Adagrad
Alexandre Défossez
Léon Bottou
Francis R. Bach
Nicolas Usunier
56
143
0
05 Mar 2020
Aggregated Residual Transformations for Deep Neural Networks
Saining Xie
Ross B. Girshick
Piotr Dollár
Z. Tu
Kaiming He
261
10,106
0
16 Nov 2016
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima
N. Keskar
Dheevatsa Mudigere
J. Nocedal
M. Smelyanskiy
P. T. P. Tang
ODL
273
2,696
0
15 Sep 2016
A Differential Equation for Modeling Nesterov's Accelerated Gradient Method: Theory and Insights
Weijie Su
Stephen P. Boyd
Emmanuel J. Candes
92
1,091
0
04 Mar 2015
1