ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1908.02246
  4. Cited By
On Convergence of Distributed Approximate Newton Methods: Globalization,
  Sharper Bounds and Beyond

On Convergence of Distributed Approximate Newton Methods: Globalization, Sharper Bounds and Beyond

Journal of machine learning research (JMLR), 2019
6 August 2019
Xiao-Tong Yuan
Ping Li
ArXiv (abs)PDFHTML

Papers citing "On Convergence of Distributed Approximate Newton Methods: Globalization, Sharper Bounds and Beyond"

21 / 21 papers shown
Accelerated Methods with Compressed Communications for Distributed
  Optimization Problems under Data Similarity
Accelerated Methods with Compressed Communications for Distributed Optimization Problems under Data SimilarityAAAI Conference on Artificial Intelligence (AAAI), 2024
Dmitry Bylinkin
Aleksandr Beznosikov
442
3
0
21 Dec 2024
Unified Gradient-Based Machine Unlearning with Remain Geometry
  Enhancement
Unified Gradient-Based Machine Unlearning with Remain Geometry EnhancementNeural Information Processing Systems (NeurIPS), 2024
Zhehao Huang
Xinwen Cheng
Jinghao Zheng
Haoran Wang
Zhengbao He
Tao Li
Xiaolin Huang
MU
299
35
0
29 Sep 2024
Accelerated Stochastic ExtraGradient: Mixing Hessian and Gradient
  Similarity to Reduce Communication in Distributed and Federated Learning
Accelerated Stochastic ExtraGradient: Mixing Hessian and Gradient Similarity to Reduce Communication in Distributed and Federated LearningУспехи математических наук (Uspekhi Mat. Nauk.), 2024
Dmitry Bylinkin
Kirill Degtyarev
Aleksandr Beznosikov
FedML
250
0
0
22 Sep 2024
A Unified Theory of Stochastic Proximal Point Methods without Smoothness
A Unified Theory of Stochastic Proximal Point Methods without Smoothness
Peter Richtárik
Abdurakhmon Sadiev
Yury Demidovich
286
8
0
24 May 2024
Federated Optimization with Doubly Regularized Drift Correction
Federated Optimization with Doubly Regularized Drift Correction
Xiaowen Jiang
Anton Rodomanov
Sebastian U. Stich
FedML
257
16
0
12 Apr 2024
Collaborative Distributed Machine Learning
Collaborative Distributed Machine LearningACM Computing Surveys (ACM Comput. Surv.), 2023
Sumit Kumar Jha
Patrick Lincoln
Sascha Rank
Ali Sunyaev
424
8
0
28 Sep 2023
Stochastic Distributed Optimization under Average Second-order
  Similarity: Algorithms and Analysis
Stochastic Distributed Optimization under Average Second-order Similarity: Algorithms and AnalysisNeural Information Processing Systems (NeurIPS), 2023
Dachao Lin
Yuze Han
Haishan Ye
Zhihua Zhang
365
15
0
15 Apr 2023
Similarity, Compression and Local Steps: Three Pillars of Efficient
  Communications for Distributed Variational Inequalities
Similarity, Compression and Local Steps: Three Pillars of Efficient Communications for Distributed Variational InequalitiesNeural Information Processing Systems (NeurIPS), 2023
Aleksandr Beznosikov
Martin Takáč
Alexander Gasnikov
346
13
0
15 Feb 2023
Sharper Analysis for Minibatch Stochastic Proximal Point Methods:
  Stability, Smoothness, and Deviation
Sharper Analysis for Minibatch Stochastic Proximal Point Methods: Stability, Smoothness, and DeviationJournal of machine learning research (JMLR), 2023
Xiao-Tong Yuan
P. Li
254
3
0
09 Jan 2023
Faster federated optimization under second-order similarity
Faster federated optimization under second-order similarityInternational Conference on Learning Representations (ICLR), 2022
Ahmed Khaled
Chi Jin
FedML
317
25
0
06 Sep 2022
Scalable K-FAC Training for Deep Neural Networks with Distributed
  Preconditioning
Scalable K-FAC Training for Deep Neural Networks with Distributed PreconditioningIEEE Transactions on Cloud Computing (IEEE TCC), 2022
Lin Zhang
Shaoshuai Shi
Wei Wang
Yue Liu
233
11
0
30 Jun 2022
Compression and Data Similarity: Combination of Two Techniques for
  Communication-Efficient Solving of Distributed Variational Inequalities
Compression and Data Similarity: Combination of Two Techniques for Communication-Efficient Solving of Distributed Variational Inequalities
Aleksandr Beznosikov
Alexander Gasnikov
274
12
0
19 Jun 2022
Optimal Gradient Sliding and its Application to Distributed Optimization
  Under Similarity
Optimal Gradient Sliding and its Application to Distributed Optimization Under Similarity
D. Kovalev
Aleksandr Beznosikov
Ekaterina Borodich
Alexander Gasnikov
G. Scutari
201
13
0
30 May 2022
Acceleration in Distributed Optimization under Similarity
Acceleration in Distributed Optimization under SimilarityInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2021
Helena Lofstrom
G. Scutari
Tianyue Cao
Alexander Gasnikov
190
31
0
24 Oct 2021
Distributed Saddle-Point Problems Under Similarity
Distributed Saddle-Point Problems Under Similarity
Aleksandr Beznosikov
G. Scutari
Alexander Rogozin
Alexander Gasnikov
453
14
0
22 Jul 2021
Robust Distributed Optimization With Randomly Corrupted Gradients
Robust Distributed Optimization With Randomly Corrupted GradientsIEEE Transactions on Signal Processing (IEEE TSP), 2021
Berkay Turan
César A. Uribe
Hoi-To Wai
M. Alizadeh
245
20
0
28 Jun 2021
Data-Free Knowledge Distillation for Heterogeneous Federated Learning
Data-Free Knowledge Distillation for Heterogeneous Federated LearningInternational Conference on Machine Learning (ICML), 2021
Zhuangdi Zhu
Junyuan Hong
Jiayu Zhou
FedML
506
931
0
20 May 2021
Newton Method over Networks is Fast up to the Statistical Precision
Newton Method over Networks is Fast up to the Statistical PrecisionInternational Conference on Machine Learning (ICML), 2021
Amir Daneshmand
G. Scutari
Pavel Dvurechensky
Alexander Gasnikov
190
22
0
12 Feb 2021
Hybrid Stochastic-Deterministic Minibatch Proximal Gradient:
  Less-Than-Single-Pass Optimization with Nearly Optimal Generalization
Hybrid Stochastic-Deterministic Minibatch Proximal Gradient: Less-Than-Single-Pass Optimization with Nearly Optimal GeneralizationInternational Conference on Machine Learning (ICML), 2020
Pan Zhou
Xiaotong Yuan
182
6
0
18 Sep 2020
Statistically Preconditioned Accelerated Gradient Method for Distributed
  Optimization
Statistically Preconditioned Accelerated Gradient Method for Distributed OptimizationInternational Conference on Machine Learning (ICML), 2020
Aymeric Dieuleveut
Lin Xiao
Sébastien Bubeck
Francis R. Bach
Laurent Massoulie
272
66
0
25 Feb 2020
Do Subsampled Newton Methods Work for High-Dimensional Data?
Do Subsampled Newton Methods Work for High-Dimensional Data?AAAI Conference on Artificial Intelligence (AAAI), 2019
Xiang Li
Shusen Wang
Zhihua Zhang
206
15
0
13 Feb 2019
1
Page 1 of 1