ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1702.06269
  4. Cited By
Memory and Communication Efficient Distributed Stochastic Optimization
  with Minibatch-Prox
v1v2 (latest)

Memory and Communication Efficient Distributed Stochastic Optimization with Minibatch-Prox

Annual Conference Computational Learning Theory (COLT), 2017
21 February 2017
Jialei Wang
Weiran Wang
Nathan Srebro
ArXiv (abs)PDFHTML

Papers citing "Memory and Communication Efficient Distributed Stochastic Optimization with Minibatch-Prox"

33 / 33 papers shown
Memory-Constrained Algorithms for Convex Optimization via Recursive
  Cutting-Planes
Memory-Constrained Algorithms for Convex Optimization via Recursive Cutting-PlanesNeural Information Processing Systems (NeurIPS), 2023
Moise Blanchard
Junhui Zhang
Patrick Jaillet
164
4
0
16 Jun 2023
Sharper Analysis for Minibatch Stochastic Proximal Point Methods:
  Stability, Smoothness, and Deviation
Sharper Analysis for Minibatch Stochastic Proximal Point Methods: Stability, Smoothness, and DeviationJournal of machine learning research (JMLR), 2023
Xiao-Tong Yuan
P. Li
225
2
0
09 Jan 2023
Uniform Stability for First-Order Empirical Risk Minimization
Uniform Stability for First-Order Empirical Risk MinimizationAnnual Conference Computational Learning Theory (COLT), 2022
Amit Attia
Tomer Koren
154
8
0
17 Jul 2022
On Convergence of FedProx: Local Dissimilarity Invariant Bounds,
  Non-smoothness and Beyond
On Convergence of FedProx: Local Dissimilarity Invariant Bounds, Non-smoothness and BeyondNeural Information Processing Systems (NeurIPS), 2022
Xiao-Tong Yuan
P. Li
FedML
194
92
0
10 Jun 2022
Minibatch and Momentum Model-based Methods for Stochastic Weakly Convex
  Optimization
Minibatch and Momentum Model-based Methods for Stochastic Weakly Convex OptimizationNeural Information Processing Systems (NeurIPS), 2021
Qi Deng
Wenzhi Gao
218
17
0
06 Jun 2021
Algorithmic Instabilities of Accelerated Gradient Descent
Algorithmic Instabilities of Accelerated Gradient DescentNeural Information Processing Systems (NeurIPS), 2021
Amit Attia
Tomer Koren
145
11
0
03 Feb 2021
The Min-Max Complexity of Distributed Stochastic Convex Optimization
  with Intermittent Communication
The Min-Max Complexity of Distributed Stochastic Convex Optimization with Intermittent CommunicationAnnual Conference Computational Learning Theory (COLT), 2021
Blake E. Woodworth
Brian Bullins
Ohad Shamir
Nathan Srebro
260
49
0
02 Feb 2021
Inverse Multiobjective Optimization Through Online Learning
Inverse Multiobjective Optimization Through Online Learning
Chaosheng Dong
Yijia Wang
Bo Zeng
131
4
0
12 Oct 2020
Adaptive Periodic Averaging: A Practical Approach to Reducing
  Communication in Distributed Learning
Adaptive Periodic Averaging: A Practical Approach to Reducing Communication in Distributed Learning
Peng Jiang
G. Agrawal
137
5
0
13 Jul 2020
Stochastic Proximal Gradient Algorithm with Minibatches. Application to
  Large Scale Learning Models
Stochastic Proximal Gradient Algorithm with Minibatches. Application to Large Scale Learning Models
A. Pătraşcu
C. Paduraru
Paul Irofti
115
0
0
30 Mar 2020
Is Local SGD Better than Minibatch SGD?
Is Local SGD Better than Minibatch SGD?International Conference on Machine Learning (ICML), 2020
Blake E. Woodworth
Kumar Kshitij Patel
Sebastian U. Stich
Zhen Dai
Brian Bullins
H. B. McMahan
Ohad Shamir
Nathan Srebro
FedML
299
271
0
18 Feb 2020
Parallel Restarted SPIDER -- Communication Efficient Distributed
  Nonconvex Optimization with Optimal Computation Complexity
Parallel Restarted SPIDER -- Communication Efficient Distributed Nonconvex Optimization with Optimal Computation Complexity
Pranay Sharma
Swatantra Kafle
Prashant Khanduri
Saikiran Bulusu
K. Rajawat
P. Varshney
FedML
285
18
0
12 Dec 2019
Least Squares Approximation for a Distributed System
Least Squares Approximation for a Distributed SystemJournal of Computational And Graphical Statistics (JCGS), 2019
Xuening Zhu
Feng Li
Hansheng Wang
394
64
0
14 Aug 2019
On Convergence of Distributed Approximate Newton Methods: Globalization,
  Sharper Bounds and Beyond
On Convergence of Distributed Approximate Newton Methods: Globalization, Sharper Bounds and BeyondJournal of machine learning research (JMLR), 2019
Xiao-Tong Yuan
Ping Li
248
35
0
06 Aug 2019
Communication-Efficient Accurate Statistical Estimation
Communication-Efficient Accurate Statistical Estimation
Jianqing Fan
Yongyi Guo
Kaizheng Wang
175
134
0
12 Jun 2019
Convergence of Distributed Stochastic Variance Reduced Methods without
  Sampling Extra Data
Convergence of Distributed Stochastic Variance Reduced Methods without Sampling Extra DataIEEE Transactions on Signal Processing (IEEE Trans. Signal Process.), 2019
Shicong Cen
Huishuai Zhang
Yuejie Chi
Wei-neng Chen
Tie-Yan Liu
FedML
225
29
0
29 May 2019
A Distributed Hierarchical SGD Algorithm with Sparse Global Reduction
A Distributed Hierarchical SGD Algorithm with Sparse Global Reduction
Fan Zhou
Guojing Cong
157
8
0
12 Mar 2019
Stochastic Approximation of Smooth and Strongly Convex Functions: Beyond
  the $O(1/T)$ Convergence Rate
Stochastic Approximation of Smooth and Strongly Convex Functions: Beyond the O(1/T)O(1/T)O(1/T) Convergence Rate
Lijun Zhang
Zhi Zhou
179
31
0
27 Jan 2019
ASVRG: Accelerated Proximal SVRG
ASVRG: Accelerated Proximal SVRG
Fanhua Shang
L. Jiao
Kaiwen Zhou
James Cheng
Yan Ren
Yufei Jin
ODL
277
35
0
07 Oct 2018
Generalized Inverse Optimization through Online Learning
Generalized Inverse Optimization through Online Learning
Chaosheng Dong
Yiran Chen
Bo Zeng
238
47
0
03 Oct 2018
Don't Use Large Mini-Batches, Use Local SGD
Don't Use Large Mini-Batches, Use Local SGD
Tao Lin
Sebastian U. Stich
Kumar Kshitij Patel
Martin Jaggi
718
454
0
22 Aug 2018
COLA: Decentralized Linear Learning
COLA: Decentralized Linear Learning
Lie He
An Bian
Martin Jaggi
244
130
0
13 Aug 2018
Robust Implicit Backpropagation
Robust Implicit Backpropagation
Francois Fagan
G. Iyengar
137
1
0
07 Aug 2018
The Effect of Network Width on the Performance of Large-batch Training
The Effect of Network Width on the Performance of Large-batch Training
Lingjiao Chen
Hongyi Wang
Jinman Zhao
Dimitris Papailiopoulos
Paraschos Koutris
195
22
0
11 Jun 2018
Graph Oracle Models, Lower Bounds, and Gaps for Parallel Stochastic
  Optimization
Graph Oracle Models, Lower Bounds, and Gaps for Parallel Stochastic Optimization
Blake E. Woodworth
Jialei Wang
Adam D. Smith
H. B. McMahan
Nathan Srebro
245
126
0
25 May 2018
Double Quantization for Communication-Efficient Distributed Optimization
Double Quantization for Communication-Efficient Distributed Optimization
Yue Yu
Jiaxiang Wu
Longbo Huang
MQ
342
59
0
25 May 2018
Distributed Stochastic Optimization via Adaptive SGD
Distributed Stochastic Optimization via Adaptive SGD
Ashok Cutkosky
R. Busa-Fekete
FedML
187
24
0
16 Feb 2018
Distributed Stochastic Multi-Task Learning with Graph Regularization
Distributed Stochastic Multi-Task Learning with Graph Regularization
Weiran Wang
Jialei Wang
Mladen Kolar
Nathan Srebro
FedML
114
21
0
11 Feb 2018
Gradient Sparsification for Communication-Efficient Distributed
  Optimization
Gradient Sparsification for Communication-Efficient Distributed OptimizationNeural Information Processing Systems (NeurIPS), 2017
Jianqiao Wangni
Jialei Wang
Ji Liu
Tong Zhang
282
572
0
26 Oct 2017
Stochastic Nonconvex Optimization with Large Minibatches
Stochastic Nonconvex Optimization with Large Minibatches
Weiran Wang
Nathan Srebro
369
27
0
25 Sep 2017
On the convergence properties of a $K$-step averaging stochastic
  gradient descent algorithm for nonconvex optimization
On the convergence properties of a KKK-step averaging stochastic gradient descent algorithm for nonconvex optimization
Fan Zhou
Guojing Cong
383
243
0
03 Aug 2017
Improved Optimization of Finite Sums with Minibatch Stochastic Variance
  Reduced Proximal Iterations
Improved Optimization of Finite Sums with Minibatch Stochastic Variance Reduced Proximal Iterations
Jialei Wang
Tong Zhang
251
12
0
21 Jun 2017
Gradient Diversity: a Key Ingredient for Scalable Distributed Learning
Gradient Diversity: a Key Ingredient for Scalable Distributed Learning
Dong Yin
A. Pananjady
Max Lam
Dimitris Papailiopoulos
Kannan Ramchandran
Peter L. Bartlett
210
12
0
18 Jun 2017
1