ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2201.08025
  4. Cited By
Low-Pass Filtering SGD for Recovering Flat Optima in the Deep Learning
  Optimization Landscape
v1v2 (latest)

Low-Pass Filtering SGD for Recovering Flat Optima in the Deep Learning Optimization Landscape

International Conference on Artificial Intelligence and Statistics (AISTATS), 2022
20 January 2022
Devansh Bisla
Jing Wang
A. Choromańska
ArXiv (abs)PDFHTML

Papers citing "Low-Pass Filtering SGD for Recovering Flat Optima in the Deep Learning Optimization Landscape"

29 / 29 papers shown
A Unified Stability Analysis of SAM vs SGD: Role of Data Coherence and Emergence of Simplicity Bias
A Unified Stability Analysis of SAM vs SGD: Role of Data Coherence and Emergence of Simplicity Bias
Wei-Kai Chang
Rajiv Khanna
MLT
232
2
0
21 Nov 2025
Stable Coresets via Posterior Sampling: Aligning Induced and Full Loss Landscapes
Stable Coresets via Posterior Sampling: Aligning Induced and Full Loss Landscapes
Wei-Kai Chang
Rajiv Khanna
221
0
0
21 Nov 2025
Flatness-Aware Stochastic Gradient Langevin Dynamics
Flatness-Aware Stochastic Gradient Langevin Dynamics
Stefano Bruno
Youngsik Hwang
Jaehyeon An
Sotirios Sabanis
Dong-Young Lim
238
0
0
02 Oct 2025
Unpacking the Implicit Norm Dynamics of Sharpness-Aware Minimization in Tensorized Models
Unpacking the Implicit Norm Dynamics of Sharpness-Aware Minimization in Tensorized Models
Tianxiao Cao
Kyohei Atarashi
H. Kashima
268
0
0
14 Aug 2025
Communication-Efficient Distributed Training for Collaborative Flat Optima Recovery in Deep Learning
Communication-Efficient Distributed Training for Collaborative Flat Optima Recovery in Deep Learning
Tolga Dimlioglu
A. Choromańska
FedML
310
1
0
27 Jul 2025
Towards Understanding The Calibration Benefits of Sharpness-Aware Minimization
Towards Understanding The Calibration Benefits of Sharpness-Aware Minimization
C. Tan
Yubo Zhou
Haishan Ye
Guang Dai
Junmin Liu
Zengjie Song
Jiangshe Zhang
Zixiang Zhao
Yunda Hao
Yong Xu
AAML
306
0
0
29 May 2025
Towards Robust Influence Functions with Flat Validation Minima
Towards Robust Influence Functions with Flat Validation MinimaInternational Conference on Machine Learning (ICML), 2025
Xichen Ye
Yifan Wu
Weizhong Zhang
Cheng Jin
Yifan Chen
TDI
476
3
0
25 May 2025
Understanding Flatness in Generative Models: Its Role and Benefits
Understanding Flatness in Generative Models: Its Role and Benefits
Taehwan Lee
Kyeongkook Seo
Jaejun Yoo
Sung Whan Yoon
DiffM
441
1
0
14 Mar 2025
Enhancing Sharpness-Aware Minimization by Learning Perturbation Radius
Enhancing Sharpness-Aware Minimization by Learning Perturbation Radius
Xuehao Wang
Weisen Jiang
Shuai Fu
Yu Zhang
AAML
290
1
0
15 Aug 2024
Enhancing Domain Adaptation through Prompt Gradient Alignment
Enhancing Domain Adaptation through Prompt Gradient Alignment
Hoang Phan
Lam C. Tran
Quyen Tran
Trung Le
683
9
0
13 Jun 2024
Revisiting Random Weight Perturbation for Efficiently Improving
  Generalization
Revisiting Random Weight Perturbation for Efficiently Improving Generalization
Tao Li
Qinghua Tao
Weihao Yan
Zehao Lei
Yingwen Wu
Kun Fang
Mingzhen He
Xiaolin Huang
AAML
434
13
0
30 Mar 2024
Friendly Sharpness-Aware Minimization
Friendly Sharpness-Aware MinimizationComputer Vision and Pattern Recognition (CVPR), 2024
Tao Li
Pan Zhou
Zhengbao He
Xinwen Cheng
Xiaolin Huang
AAML
288
43
0
19 Mar 2024
GRAWA: Gradient-based Weighted Averaging for Distributed Training of
  Deep Learning Models
GRAWA: Gradient-based Weighted Averaging for Distributed Training of Deep Learning Models
Tolga Dimlioglu
A. Choromańska
255
6
0
07 Mar 2024
Stabilizing Sharpness-aware Minimization Through A Simple
  Renormalization Strategy
Stabilizing Sharpness-aware Minimization Through A Simple Renormalization Strategy
Chengli Tan
Jiangshe Zhang
Junmin Liu
Yicheng Wang
Yunda Hao
AAML
342
8
0
14 Jan 2024
Selectivity Drives Productivity: Efficient Dataset Pruning for Enhanced
  Transfer Learning
Selectivity Drives Productivity: Efficient Dataset Pruning for Enhanced Transfer LearningNeural Information Processing Systems (NeurIPS), 2023
Yihua Zhang
Yimeng Zhang
Chenyi Zi
Jinghan Jia
Jiancheng Liu
Gaowen Liu
Min-Fong Hong
Shiyu Chang
Sijia Liu
AAML
412
15
0
13 Oct 2023
Based on What We Can Control Artificial Neural Networks
Based on What We Can Control Artificial Neural Networks
Cheng Kang
Xujing Yao
210
0
0
09 Oct 2023
Entropy-MCMC: Sampling from Flat Basins with Ease
Entropy-MCMC: Sampling from Flat Basins with EaseInternational Conference on Learning Representations (ICLR), 2023
Bolian Li
Ruqi Zhang
636
7
0
09 Oct 2023
Decentralized SGD and Average-direction SAM are Asymptotically
  Equivalent
Decentralized SGD and Average-direction SAM are Asymptotically EquivalentInternational Conference on Machine Learning (ICML), 2023
Tongtian Zhu
Fengxiang He
Kaixuan Chen
Weilong Dai
Dacheng Tao
721
20
0
05 Jun 2023
An Adaptive Policy to Employ Sharpness-Aware Minimization
An Adaptive Policy to Employ Sharpness-Aware MinimizationInternational Conference on Learning Representations (ICLR), 2023
Weisen Jiang
Hansi Yang
Yu Zhang
James T. Kwok
AAML
297
46
0
28 Apr 2023
Going Further: Flatness at the Rescue of Early Stopping for Adversarial
  Example Transferability
Going Further: Flatness at the Rescue of Early Stopping for Adversarial Example Transferability
Martin Gubri
Maxime Cordy
Yves Le Traon
AAML
267
3
1
05 Apr 2023
A Modern Look at the Relationship between Sharpness and Generalization
A Modern Look at the Relationship between Sharpness and GeneralizationInternational Conference on Machine Learning (ICML), 2023
Maksym Andriushchenko
Francesco Croce
Maximilian Müller
Matthias Hein
Nicolas Flammarion
3DH
356
90
0
14 Feb 2023
Escaping Saddle Points for Effective Generalization on Class-Imbalanced
  Data
Escaping Saddle Points for Effective Generalization on Class-Imbalanced DataNeural Information Processing Systems (NeurIPS), 2022
Harsh Rangwani
Sumukh K Aithal
Mayank Mishra
R. Venkatesh Babu
252
41
0
28 Dec 2022
Improving Generalization of Pre-trained Language Models via Stochastic
  Weight Averaging
Improving Generalization of Pre-trained Language Models via Stochastic Weight AveragingConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Peng Lu
I. Kobyzev
Mehdi Rezagholizadeh
Ahmad Rashid
A. Ghodsi
Philippe Langlais
MoMe
250
12
0
12 Dec 2022
Efficient Generalization Improvement Guided by Random Weight
  Perturbation
Efficient Generalization Improvement Guided by Random Weight Perturbation
Tao Li
Wei Yan
Zehao Lei
Yingwen Wu
Kun Fang
Ming-Hsuan Yang
Xiaolin Huang
AAML
174
8
0
21 Nov 2022
SAM as an Optimal Relaxation of Bayes
SAM as an Optimal Relaxation of BayesInternational Conference on Learning Representations (ICLR), 2022
Thomas Möllenhoff
Mohammad Emtiyaz Khan
BDL
360
42
0
04 Oct 2022
Trainable Weight Averaging: Accelerating Training and Improving Generalization
Trainable Weight Averaging: Accelerating Training and Improving Generalization
Tao Li
Zhehao Huang
Yingwen Wu
Zhengbao He
Qinghua Tao
Xiaolin Huang
Chih-Jen Lin
MoMe
377
3
0
26 May 2022
Train Flat, Then Compress: Sharpness-Aware Minimization Learns More
  Compressible Models
Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Clara Na
Sanket Vaibhav Mehta
Emma Strubell
315
24
0
25 May 2022
Anticorrelated Noise Injection for Improved Generalization
Anticorrelated Noise Injection for Improved GeneralizationInternational Conference on Machine Learning (ICML), 2022
Antonio Orvieto
Hans Kersting
F. Proske
Francis R. Bach
Aurelien Lucchi
375
58
0
06 Feb 2022
When Do Flat Minima Optimizers Work?
When Do Flat Minima Optimizers Work?Neural Information Processing Systems (NeurIPS), 2022
Jean Kaddour
Linqing Liu
Ricardo M. A. Silva
Matt J. Kusner
ODL
596
91
0
01 Feb 2022
1
Page 1 of 1