ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2007.13985
  4. Cited By
Stochastic Normalized Gradient Descent with Momentum for Large-Batch
  Training
v1v2 (latest)

Stochastic Normalized Gradient Descent with Momentum for Large-Batch Training

Science China Information Sciences (Sci China Inf Sci), 2020
28 July 2020
Shen-Yi Zhao
Chang-Wei Shi
Yin-Peng Xie
Wu-Jun Li
    ODL
ArXiv (abs)PDFHTML

Papers citing "Stochastic Normalized Gradient Descent with Momentum for Large-Batch Training"

7 / 7 papers shown
AdamNX: An Adam improvement algorithm based on a novel exponential decay mechanism for the second-order moment estimate
AdamNX: An Adam improvement algorithm based on a novel exponential decay mechanism for the second-order moment estimate
Meng Zhu
Quan Xiao
Weidong Min
319
0
0
17 Nov 2025
Neural Optimizer Equation, Decay Function, and Learning Rate Schedule
  Joint Evolution
Neural Optimizer Equation, Decay Function, and Learning Rate Schedule Joint EvolutionAnnual Conference on Genetic and Evolutionary Computation (GECCO), 2024
Brandon Morgan
Dean Frederick Hougen
ODL
317
0
0
10 Apr 2024
On the Optimal Batch Size for Byzantine-Robust Distributed Learning
On the Optimal Batch Size for Byzantine-Robust Distributed Learning
Yi-Rui Yang
Chang-Wei Shi
Wu-Jun Li
FedMLAAML
314
1
0
23 May 2023
Revisiting Outer Optimization in Adversarial Training
Revisiting Outer Optimization in Adversarial TrainingEuropean Conference on Computer Vision (ECCV), 2022
Ali Dabouei
Fariborz Taherkhani
Sobhan Soleymani
Nasser M. Nasrabadi
AAML
305
6
0
02 Sep 2022
Automatic Clipping: Differentially Private Deep Learning Made Easier and
  Stronger
Automatic Clipping: Differentially Private Deep Learning Made Easier and StrongerNeural Information Processing Systems (NeurIPS), 2022
Zhiqi Bu
Yu Wang
Sheng Zha
George Karypis
837
111
0
14 Jun 2022
Descending through a Crowded Valley - Benchmarking Deep Learning
  Optimizers
Descending through a Crowded Valley - Benchmarking Deep Learning Optimizers
Robin M. Schmidt
Frank Schneider
Philipp Hennig
ODL
919
195
0
03 Jul 2020
Global Momentum Compression for Sparse Communication in Distributed
  Learning
Global Momentum Compression for Sparse Communication in Distributed Learning
Chang-Wei Shi
Shen-Yi Zhao
Yin-Peng Xie
Hao Gao
Wu-Jun Li
406
1
0
30 May 2019
1
Page 1 of 1