Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1609.04836
Cited By
v1
v2 (latest)
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima
15 September 2016
N. Keskar
Dheevatsa Mudigere
J. Nocedal
M. Smelyanskiy
P. T. P. Tang
ODL
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima"
4 / 1,554 papers shown
Title
Distributed Training of Deep Neural Networks: Theoretical and Practical Limits of Parallel Scalability
J. Keuper
Franz-Josef Pfreundt
GNN
119
98
0
22 Sep 2016
Spectral Methods for Correlated Topic Models
Forough Arabshahi
Anima Anandkumar
OOD
45
2
0
30 May 2016
Parallelizing Word2Vec in Shared and Distributed Memory
Shihao Ji
N. Satish
Sheng Li
Pradeep Dubey
VLM
MoE
51
72
0
15 Apr 2016
Revisiting Distributed Synchronous SGD
Jianmin Chen
Xinghao Pan
R. Monga
Samy Bengio
Rafal Jozefowicz
91
801
0
04 Apr 2016
Previous
1
2
3
...
30
31
32