ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2201.08025
  4. Cited By
Low-Pass Filtering SGD for Recovering Flat Optima in the Deep Learning
  Optimization Landscape
v1v2 (latest)

Low-Pass Filtering SGD for Recovering Flat Optima in the Deep Learning Optimization Landscape

International Conference on Artificial Intelligence and Statistics (AISTATS), 2022
20 January 2022
Devansh Bisla
Jing Wang
A. Choromańska
ArXiv (abs)PDFHTML

Papers citing "Low-Pass Filtering SGD for Recovering Flat Optima in the Deep Learning Optimization Landscape"

29 / 29 papers shown
A Unified Stability Analysis of SAM vs SGD: Role of Data Coherence and Emergence of Simplicity Bias
A Unified Stability Analysis of SAM vs SGD: Role of Data Coherence and Emergence of Simplicity Bias
Wei-Kai Chang
Rajiv Khanna
MLT
215
2
0
21 Nov 2025
Stable Coresets via Posterior Sampling: Aligning Induced and Full Loss Landscapes
Stable Coresets via Posterior Sampling: Aligning Induced and Full Loss Landscapes
Wei-Kai Chang
Rajiv Khanna
210
0
0
21 Nov 2025
Flatness-Aware Stochastic Gradient Langevin Dynamics
Flatness-Aware Stochastic Gradient Langevin Dynamics
Stefano Bruno
Youngsik Hwang
Jaehyeon An
Sotirios Sabanis
Dong-Young Lim
212
0
0
02 Oct 2025
Unpacking the Implicit Norm Dynamics of Sharpness-Aware Minimization in Tensorized Models
Unpacking the Implicit Norm Dynamics of Sharpness-Aware Minimization in Tensorized Models
Tianxiao Cao
Kyohei Atarashi
H. Kashima
256
0
0
14 Aug 2025
Communication-Efficient Distributed Training for Collaborative Flat Optima Recovery in Deep Learning
Communication-Efficient Distributed Training for Collaborative Flat Optima Recovery in Deep Learning
Tolga Dimlioglu
A. Choromańska
FedML
299
1
0
27 Jul 2025
Towards Understanding The Calibration Benefits of Sharpness-Aware Minimization
Towards Understanding The Calibration Benefits of Sharpness-Aware Minimization
C. Tan
Yubo Zhou
Haishan Ye
Guang Dai
Junmin Liu
Zengjie Song
Jiangshe Zhang
Zixiang Zhao
Yunda Hao
Yong Xu
AAML
299
0
0
29 May 2025
Towards Robust Influence Functions with Flat Validation Minima
Towards Robust Influence Functions with Flat Validation MinimaInternational Conference on Machine Learning (ICML), 2025
Xichen Ye
Yifan Wu
Weizhong Zhang
Cheng Jin
Yifan Chen
TDI
412
3
0
25 May 2025
Understanding Flatness in Generative Models: Its Role and Benefits
Understanding Flatness in Generative Models: Its Role and Benefits
Taehwan Lee
Kyeongkook Seo
Jaejun Yoo
Sung Whan Yoon
DiffM
425
1
0
14 Mar 2025
Enhancing Sharpness-Aware Minimization by Learning Perturbation Radius
Enhancing Sharpness-Aware Minimization by Learning Perturbation Radius
Xuehao Wang
Weisen Jiang
Shuai Fu
Yu Zhang
AAML
264
1
0
15 Aug 2024
Enhancing Domain Adaptation through Prompt Gradient Alignment
Enhancing Domain Adaptation through Prompt Gradient Alignment
Hoang Phan
Lam C. Tran
Quyen Tran
Trung Le
636
8
0
13 Jun 2024
Revisiting Random Weight Perturbation for Efficiently Improving
  Generalization
Revisiting Random Weight Perturbation for Efficiently Improving Generalization
Tao Li
Qinghua Tao
Weihao Yan
Zehao Lei
Yingwen Wu
Kun Fang
Mingzhen He
Xiaolin Huang
AAML
419
13
0
30 Mar 2024
Friendly Sharpness-Aware Minimization
Friendly Sharpness-Aware MinimizationComputer Vision and Pattern Recognition (CVPR), 2024
Tao Li
Pan Zhou
Zhengbao He
Xinwen Cheng
Xiaolin Huang
AAML
275
40
0
19 Mar 2024
GRAWA: Gradient-based Weighted Averaging for Distributed Training of
  Deep Learning Models
GRAWA: Gradient-based Weighted Averaging for Distributed Training of Deep Learning Models
Tolga Dimlioglu
A. Choromańska
248
6
0
07 Mar 2024
Stabilizing Sharpness-aware Minimization Through A Simple
  Renormalization Strategy
Stabilizing Sharpness-aware Minimization Through A Simple Renormalization Strategy
Chengli Tan
Jiangshe Zhang
Junmin Liu
Yicheng Wang
Yunda Hao
AAML
332
7
0
14 Jan 2024
Selectivity Drives Productivity: Efficient Dataset Pruning for Enhanced
  Transfer Learning
Selectivity Drives Productivity: Efficient Dataset Pruning for Enhanced Transfer LearningNeural Information Processing Systems (NeurIPS), 2023
Yihua Zhang
Yimeng Zhang
Chenyi Zi
Jinghan Jia
Jiancheng Liu
Gaowen Liu
Min-Fong Hong
Shiyu Chang
Sijia Liu
AAML
396
14
0
13 Oct 2023
Based on What We Can Control Artificial Neural Networks
Based on What We Can Control Artificial Neural Networks
Cheng Kang
Xujing Yao
205
0
0
09 Oct 2023
Entropy-MCMC: Sampling from Flat Basins with Ease
Entropy-MCMC: Sampling from Flat Basins with EaseInternational Conference on Learning Representations (ICLR), 2023
Bolian Li
Ruqi Zhang
604
7
0
09 Oct 2023
Decentralized SGD and Average-direction SAM are Asymptotically
  Equivalent
Decentralized SGD and Average-direction SAM are Asymptotically EquivalentInternational Conference on Machine Learning (ICML), 2023
Tongtian Zhu
Fengxiang He
Kaixuan Chen
Weilong Dai
Dacheng Tao
703
20
0
05 Jun 2023
An Adaptive Policy to Employ Sharpness-Aware Minimization
An Adaptive Policy to Employ Sharpness-Aware MinimizationInternational Conference on Learning Representations (ICLR), 2023
Weisen Jiang
Hansi Yang
Yu Zhang
James T. Kwok
AAML
289
44
0
28 Apr 2023
Going Further: Flatness at the Rescue of Early Stopping for Adversarial
  Example Transferability
Going Further: Flatness at the Rescue of Early Stopping for Adversarial Example Transferability
Martin Gubri
Maxime Cordy
Yves Le Traon
AAML
254
3
1
05 Apr 2023
A Modern Look at the Relationship between Sharpness and Generalization
A Modern Look at the Relationship between Sharpness and GeneralizationInternational Conference on Machine Learning (ICML), 2023
Maksym Andriushchenko
Francesco Croce
Maximilian Müller
Matthias Hein
Nicolas Flammarion
3DH
346
86
0
14 Feb 2023
Escaping Saddle Points for Effective Generalization on Class-Imbalanced
  Data
Escaping Saddle Points for Effective Generalization on Class-Imbalanced DataNeural Information Processing Systems (NeurIPS), 2022
Harsh Rangwani
Sumukh K Aithal
Mayank Mishra
R. Venkatesh Babu
228
39
0
28 Dec 2022
Improving Generalization of Pre-trained Language Models via Stochastic
  Weight Averaging
Improving Generalization of Pre-trained Language Models via Stochastic Weight AveragingConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Peng Lu
I. Kobyzev
Mehdi Rezagholizadeh
Ahmad Rashid
A. Ghodsi
Philippe Langlais
MoMe
225
12
0
12 Dec 2022
Efficient Generalization Improvement Guided by Random Weight
  Perturbation
Efficient Generalization Improvement Guided by Random Weight Perturbation
Tao Li
Wei Yan
Zehao Lei
Yingwen Wu
Kun Fang
Ming-Hsuan Yang
Xiaolin Huang
AAML
149
8
0
21 Nov 2022
SAM as an Optimal Relaxation of Bayes
SAM as an Optimal Relaxation of BayesInternational Conference on Learning Representations (ICLR), 2022
Thomas Möllenhoff
Mohammad Emtiyaz Khan
BDL
301
41
0
04 Oct 2022
Trainable Weight Averaging: Accelerating Training and Improving Generalization
Trainable Weight Averaging: Accelerating Training and Improving Generalization
Tao Li
Zhehao Huang
Yingwen Wu
Zhengbao He
Qinghua Tao
Xiaolin Huang
Chih-Jen Lin
MoMe
349
3
0
26 May 2022
Train Flat, Then Compress: Sharpness-Aware Minimization Learns More
  Compressible Models
Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Clara Na
Sanket Vaibhav Mehta
Emma Strubell
301
24
0
25 May 2022
Anticorrelated Noise Injection for Improved Generalization
Anticorrelated Noise Injection for Improved GeneralizationInternational Conference on Machine Learning (ICML), 2022
Antonio Orvieto
Hans Kersting
F. Proske
Francis R. Bach
Aurelien Lucchi
319
57
0
06 Feb 2022
When Do Flat Minima Optimizers Work?
When Do Flat Minima Optimizers Work?Neural Information Processing Systems (NeurIPS), 2022
Jean Kaddour
Linqing Liu
Ricardo M. A. Silva
Matt J. Kusner
ODL
550
89
0
01 Feb 2022
1
Page 1 of 1