ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1808.06296
  4. Cited By
Universal Stagewise Learning for Non-Convex Problems with Convergence on
  Averaged Solutions
v1v2v3 (latest)

Universal Stagewise Learning for Non-Convex Problems with Convergence on Averaged Solutions

20 August 2018
Zaiyi Chen
Zhuoning Yuan
Jinfeng Yi
Bowen Zhou
Enhong Chen
Tianbao Yang
ArXiv (abs)PDFHTML

Papers citing "Universal Stagewise Learning for Non-Convex Problems with Convergence on Averaged Solutions"

28 / 28 papers shown
Obtaining Lower Query Complexities through Lightweight Zeroth-Order
  Proximal Gradient Algorithms
Obtaining Lower Query Complexities through Lightweight Zeroth-Order Proximal Gradient AlgorithmsNeural Computation (Neural Comput.), 2024
Bin Gu
Xiyuan Wei
Hualin Zhang
Yi Chang
Heng-Chiao Huang
FedML
237
0
0
03 Oct 2024
Differentially Private Zeroth-Order Methods for Scalable Large Language
  Model Finetuning
Differentially Private Zeroth-Order Methods for Scalable Large Language Model Finetuning
Zhicheng Liu
Jian Lou
Wenxuan Bao
Yihan Hu
Baochun Li
Zhan Qin
K. Ren
525
15
0
12 Feb 2024
Robustness to Unbounded Smoothness of Generalized SignSGD
Robustness to Unbounded Smoothness of Generalized SignSGDNeural Information Processing Systems (NeurIPS), 2022
M. Crawshaw
Mingrui Liu
Francesco Orabona
Wei Zhang
Zhenxun Zhuang
AAML
377
95
0
23 Aug 2022
When AUC meets DRO: Optimizing Partial AUC for Deep Learning with
  Non-Convex Convergence Guarantee
When AUC meets DRO: Optimizing Partial AUC for Deep Learning with Non-Convex Convergence GuaranteeInternational Conference on Machine Learning (ICML), 2022
Dixian Zhu
Gang Li
Bokun Wang
Xiaodong Wu
Tianbao Yang
712
41
0
01 Mar 2022
Toward Communication Efficient Adaptive Gradient Method
Toward Communication Efficient Adaptive Gradient Method
Xiangyi Chen
Xiaoyun Li
P. Li
FedML
220
50
0
10 Sep 2021
Local AdaGrad-Type Algorithm for Stochastic Convex-Concave Optimization
Local AdaGrad-Type Algorithm for Stochastic Convex-Concave OptimizationMachine-mediated learning (ML), 2021
Luofeng Liao
Li Shen
Jia Duan
Mladen Kolar
Dacheng Tao
291
5
0
18 Jun 2021
Bandwidth-based Step-Sizes for Non-Convex Stochastic Optimization
Bandwidth-based Step-Sizes for Non-Convex Stochastic Optimization
Xiaoyu Wang
M. Johansson
293
2
0
05 Jun 2021
On the Convergence of Step Decay Step-Size for Stochastic Optimization
On the Convergence of Step Decay Step-Size for Stochastic OptimizationNeural Information Processing Systems (NeurIPS), 2021
Xiaoyu Wang
Sindri Magnússon
M. Johansson
278
31
0
18 Feb 2021
Towards Practical Adam: Non-Convexity, Convergence Theory, and
  Mini-Batch Acceleration
Towards Practical Adam: Non-Convexity, Convergence Theory, and Mini-Batch AccelerationJournal of machine learning research (JMLR), 2021
Congliang Chen
Li Shen
Fangyu Zou
Wei Liu
240
42
0
14 Jan 2021
Adam$^+$: A Stochastic Method with Adaptive Variance Reduction
Adam+^++: A Stochastic Method with Adaptive Variance Reduction
Mingrui Liu
Wei Zhang
Francesco Orabona
Tianbao Yang
213
33
0
24 Nov 2020
Effective Federated Adaptive Gradient Methods with Non-IID Decentralized
  Data
Effective Federated Adaptive Gradient Methods with Non-IID Decentralized Data
Qianqian Tong
Guannan Liang
J. Bi
FedML
351
28
0
14 Sep 2020
Faster Stochastic Alternating Direction Method of Multipliers for
  Nonconvex Optimization
Faster Stochastic Alternating Direction Method of Multipliers for Nonconvex Optimization
Feihu Huang
Songcan Chen
Heng-Chiao Huang
324
42
0
04 Aug 2020
An Online Method for A Class of Distributionally Robust Optimization
  with Non-Convex Objectives
An Online Method for A Class of Distributionally Robust Optimization with Non-Convex Objectives
Qi Qi
Zhishuai Guo
Yi Tian Xu
Rong Jin
Tianbao Yang
486
55
0
17 Jun 2020
STL-SGD: Speeding Up Local SGD with Stagewise Communication Period
STL-SGD: Speeding Up Local SGD with Stagewise Communication PeriodAAAI Conference on Artificial Intelligence (AAAI), 2020
Shuheng Shen
Yifei Cheng
Jingchang Liu
Linli Xu
LRM
311
12
0
11 Jun 2020
Revisiting SGD with Increasingly Weighted Averaging: Optimization and
  Generalization Perspectives
Revisiting SGD with Increasingly Weighted Averaging: Optimization and Generalization Perspectives
Zhishuai Guo
Yan Yan
Tianbao Yang
MoMe
321
4
0
09 Mar 2020
Domain Decluttering: Simplifying Images to Mitigate Synthetic-Real
  Domain Shift and Improve Depth Estimation
Domain Decluttering: Simplifying Images to Mitigate Synthetic-Real Domain Shift and Improve Depth EstimationComputer Vision and Pattern Recognition (CVPR), 2020
Yunhan Zhao
Shu Kong
Daeyun Shin
Charless C. Fowlkes
MDE
261
47
0
27 Feb 2020
Stagewise Enlargement of Batch Size for SGD-based Learning
Stagewise Enlargement of Batch Size for SGD-based Learning
Shen-Yi Zhao
Yin-Peng Xie
Wu-Jun Li
174
1
0
26 Feb 2020
Biased Stochastic First-Order Methods for Conditional Stochastic
  Optimization and Applications in Meta Learning
Biased Stochastic First-Order Methods for Conditional Stochastic Optimization and Applications in Meta LearningNeural Information Processing Systems (NeurIPS), 2020
Yifan Hu
Siqi Zhang
Xin Chen
Niao He
ODL
251
62
0
25 Feb 2020
Adaptive Temporal Difference Learning with Linear Function Approximation
Adaptive Temporal Difference Learning with Linear Function ApproximationIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2020
Tao Sun
Han Shen
Tianyi Chen
Dongsheng Li
191
25
0
20 Feb 2020
Stochastic AUC Maximization with Deep Neural Networks
Stochastic AUC Maximization with Deep Neural NetworksInternational Conference on Learning Representations (ICLR), 2019
Mingrui Liu
Zhuoning Yuan
Yiming Ying
Tianbao Yang
580
116
0
28 Aug 2019
Stochastic Optimization for Non-convex Inf-Projection Problems
Stochastic Optimization for Non-convex Inf-Projection ProblemsInternational Conference on Machine Learning (ICML), 2019
Yan Yan
Yi Tian Xu
Lijun Zhang
Xiaoyu Wang
Tianbao Yang
196
3
0
26 Aug 2019
ADASS: Adaptive Sample Selection for Training Acceleration
ADASS: Adaptive Sample Selection for Training Acceleration
Shen-Yi Zhao
Hao Gao
Wu-Jun Li
323
0
0
11 Jun 2019
On the Convergence of Memory-Based Distributed SGD
On the Convergence of Memory-Based Distributed SGD
Shen-Yi Zhao
Hao Gao
Wu-Jun Li
137
1
0
30 May 2019
An Optimistic Acceleration of AMSGrad for Nonconvex Optimization
An Optimistic Acceleration of AMSGrad for Nonconvex Optimization
Jun-Kun Wang
Xiaoyun Li
Belhal Karimi
Ping Li
ODL
384
1
0
04 Mar 2019
Stagewise Training Accelerates Convergence of Testing Error Over SGD
Stagewise Training Accelerates Convergence of Testing Error Over SGD
Zhuoning Yuan
Yan Yan
Rong Jin
Tianbao Yang
300
11
0
10 Dec 2018
Stochastic Optimization for DC Functions and Non-smooth Non-convex
  Regularizers with Non-asymptotic Convergence
Stochastic Optimization for DC Functions and Non-smooth Non-convex Regularizers with Non-asymptotic Convergence
Yi Tian Xu
Qi Qi
Qihang Lin
Rong Jin
Tianbao Yang
275
48
0
28 Nov 2018
A Sufficient Condition for Convergences of Adam and RMSProp
A Sufficient Condition for Convergences of Adam and RMSPropComputer Vision and Pattern Recognition (CVPR), 2018
Fangyu Zou
Li Shen
Zequn Jie
Weizhong Zhang
Wei Liu
365
438
0
23 Nov 2018
Weakly-Convex Concave Min-Max Optimization: Provable Algorithms and
  Applications in Machine Learning
Weakly-Convex Concave Min-Max Optimization: Provable Algorithms and Applications in Machine Learning
Hassan Rafique
Mingrui Liu
Qihang Lin
Tianbao Yang
582
129
0
04 Oct 2018
1
Page 1 of 1