ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1810.12281
  4. Cited By
Three Mechanisms of Weight Decay Regularization

Three Mechanisms of Weight Decay Regularization

29 October 2018
Guodong Zhang
Chaoqi Wang
Bowen Xu
Roger C. Grosse
ArXivPDFHTML

Papers citing "Three Mechanisms of Weight Decay Regularization"

8 / 58 papers shown
Title
Gradient Centralization: A New Optimization Technique for Deep Neural
  Networks
Gradient Centralization: A New Optimization Technique for Deep Neural Networks
Hongwei Yong
Jianqiang Huang
Xiansheng Hua
Lei Zhang
ODL
32
184
0
03 Apr 2020
Highly Efficient Salient Object Detection with 100K Parameters
Highly Efficient Salient Object Detection with 100K Parameters
Shanghua Gao
Yong-qiang Tan
Ming-Ming Cheng
Chengze Lu
Yunpeng Chen
Shuicheng Yan
231
168
0
12 Mar 2020
Iterative Averaging in the Quest for Best Test Error
Iterative Averaging in the Quest for Best Test Error
Diego Granziol
Xingchen Wan
Samuel Albanie
Stephen J. Roberts
32
3
0
02 Mar 2020
Topologically Densified Distributions
Topologically Densified Distributions
Christoph Hofer
Florian Graf
Marc Niethammer
Roland Kwitt
27
15
0
12 Feb 2020
FixMatch: Simplifying Semi-Supervised Learning with Consistency and
  Confidence
FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence
Kihyuk Sohn
David Berthelot
Chun-Liang Li
Zizhao Zhang
Nicholas Carlini
E. D. Cubuk
Alexey Kurakin
Han Zhang
Colin Raffel
AAML
104
3,479
0
21 Jan 2020
An Empirical Study of Large-Batch Stochastic Gradient Descent with
  Structured Covariance Noise
An Empirical Study of Large-Batch Stochastic Gradient Descent with Structured Covariance Noise
Yeming Wen
Kevin Luk
Maxime Gazeau
Guodong Zhang
Harris Chan
Jimmy Ba
ODL
25
22
0
21 Feb 2019
Do Better ImageNet Models Transfer Better?
Do Better ImageNet Models Transfer Better?
Simon Kornblith
Jonathon Shlens
Quoc V. Le
OOD
MLT
91
1,314
0
23 May 2018
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp
  Minima
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima
N. Keskar
Dheevatsa Mudigere
J. Nocedal
M. Smelyanskiy
P. T. P. Tang
ODL
314
2,900
0
15 Sep 2016
Previous
12