ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1903.09321
  4. Cited By
WONDER: Weighted one-shot distributed ridge regression in high
  dimensions
v1v2 (latest)

WONDER: Weighted one-shot distributed ridge regression in high dimensions

International Conference on Machine Learning (ICML), 2019
22 March 2019
Guang Cheng
Yueqi Sheng
    OffRL
ArXiv (abs)PDFHTML

Papers citing "WONDER: Weighted one-shot distributed ridge regression in high dimensions"

30 / 30 papers shown
Transfer learning via Regularized Linear Discriminant Analysis
Transfer learning via Regularized Linear Discriminant Analysis
Hongzhe Zhang
Arnab Auddy
Hongzhe Lee
431
0
0
05 Jan 2025
Recent and Upcoming Developments in Randomized Numerical Linear Algebra
  for Machine Learning
Recent and Upcoming Developments in Randomized Numerical Linear Algebra for Machine Learning
Michał Dereziński
Michael W. Mahoney
333
22
0
17 Jun 2024
Asymptotically free sketched ridge ensembles: Risks, cross-validation,
  and tuning
Asymptotically free sketched ridge ensembles: Risks, cross-validation, and tuningInternational Conference on Learning Representations (ICLR), 2023
Filip Szatkowski
Daniel LeJeune
524
11
0
06 Oct 2023
On High-Dimensional Asymptotic Properties of Model Averaging Estimators
On High-Dimensional Asymptotic Properties of Model Averaging Estimators
Ryo Ando
F. Komaki
MoMe
138
6
0
18 Aug 2023
Transfer Learning with Random Coefficient Ridge Regression
Transfer Learning with Random Coefficient Ridge Regression
Hongzhe Zhang
Hongzhe Li
181
2
0
28 Jun 2023
Batches Stabilize the Minimum Norm Risk in High Dimensional
  Overparameterized Linear Regression
Batches Stabilize the Minimum Norm Risk in High Dimensional Overparameterized Linear Regression
Shahar Stein Ioushua
Inbar Hasidim
O. Shayevitz
M. Feder
332
1
0
14 Jun 2023
Generalized equivalences between subsampling and ridge regularization
Generalized equivalences between subsampling and ridge regularizationNeural Information Processing Systems (NeurIPS), 2023
Pratik V. Patil
Jin-Hong Du
387
6
0
29 May 2023
Demystifying Disagreement-on-the-Line in High Dimensions
Demystifying Disagreement-on-the-Line in High DimensionsInternational Conference on Machine Learning (ICML), 2023
Dong-Hwan Lee
Behrad Moniri
Xinmeng Huang
Guang Cheng
Hamed Hassani
400
12
0
31 Jan 2023
Distributed Sparse Linear Regression under Communication Constraints
Distributed Sparse Linear Regression under Communication Constraints
R. Fonseca
B. Nadler
FedML
443
2
0
09 Jan 2023
Asymptotics of the Sketched Pseudoinverse
Asymptotics of the Sketched PseudoinverseSIAM Journal on Mathematics of Data Science (SIMODS), 2022
Daniel LeJeune
Pratik V. Patil
Hamid Javadi
Richard G. Baraniuk
Robert Tibshirani
291
11
0
07 Nov 2022
Local SGD in Overparameterized Linear Regression
Local SGD in Overparameterized Linear Regression
Mike Nguyen
Charly Kirst
Nicole Mücke
185
0
0
20 Oct 2022
Federated Coordinate Descent for Privacy-Preserving Multiparty Linear
  Regression
Federated Coordinate Descent for Privacy-Preserving Multiparty Linear Regression
Xinlin Leng
Chenxu Li
Weifeng Xu
Yuyan Sun
Hongtao Wang
FedML
351
1
0
16 Sep 2022
Information bottleneck theory of high-dimensional regression: relevancy,
  efficiency and optimality
Information bottleneck theory of high-dimensional regression: relevancy, efficiency and optimalityNeural Information Processing Systems (NeurIPS), 2022
Wave Ngampruetikorn
David J. Schwab
216
10
0
08 Aug 2022
ReBoot: Distributed statistical learning via refitting bootstrap samples
ReBoot: Distributed statistical learning via refitting bootstrap samples
Yumeng Wang
Ziwei Zhu
Xuming He
FedMLBDL
350
1
0
19 Jul 2022
On Optimal Early Stopping: Over-informative versus Under-informative
  Parametrization
On Optimal Early Stopping: Over-informative versus Under-informative Parametrization
Ruoqi Shen
Liyao (Mars) Gao
Yi-An Ma
351
16
0
20 Feb 2022
Parameters or Privacy: A Provable Tradeoff Between Overparameterization
  and Membership Inference
Parameters or Privacy: A Provable Tradeoff Between Overparameterization and Membership InferenceNeural Information Processing Systems (NeurIPS), 2022
Jasper Tan
Blake Mason
Hamid Javadi
Richard G. Baraniuk
FedML
332
21
0
02 Feb 2022
Data splitting improves statistical performance in overparametrized
  regimes
Data splitting improves statistical performance in overparametrized regimesInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2021
Nicole Mücke
Enrico Reiss
Jonas Rungenhagen
Markus Klein
237
8
0
21 Oct 2021
Solon: Communication-efficient Byzantine-resilient Distributed Training
  via Redundant Gradients
Solon: Communication-efficient Byzantine-resilient Distributed Training via Redundant Gradients
Lingjiao Chen
Leshang Chen
Hongyi Wang
S. Davidson
Guang Cheng
FedML
340
1
0
04 Oct 2021
The Common Intuition to Transfer Learning Can Win or Lose: Case Studies
  for Linear Regression
The Common Intuition to Transfer Learning Can Win or Lose: Case Studies for Linear RegressionSIAM Journal on Mathematics of Data Science (SIMODS), 2021
Yehuda Dar
Daniel LeJeune
Richard G. Baraniuk
MLT
270
8
0
09 Mar 2021
Sparse sketches with small inversion bias
Sparse sketches with small inversion biasAnnual Conference Computational Learning Theory (COLT), 2020
Michal Derezinski
Zhenyu Liao
Guang Cheng
Michael W. Mahoney
515
25
0
21 Nov 2020
$σ$-Ridge: group regularized ridge regression via empirical Bayes
  noise level cross-validation
σσσ-Ridge: group regularized ridge regression via empirical Bayes noise level cross-validation
Nikolaos Ignatiadis
Panagiotis Lolas
279
5
0
29 Oct 2020
Precise High-Dimensional Asymptotics for Quantifying Heterogeneous Transfers
Precise High-Dimensional Asymptotics for Quantifying Heterogeneous Transfers
Fan Yang
Hongyang R. Zhang
Sen Wu
Christopher Ré
Weijie J. Su
561
24
0
22 Oct 2020
What causes the test error? Going beyond bias-variance via ANOVA
What causes the test error? Going beyond bias-variance via ANOVAJournal of machine learning research (JMLR), 2020
Licong Lin
Guang Cheng
328
36
0
11 Oct 2020
Canonical thresholding for non-sparse high-dimensional linear regression
Canonical thresholding for non-sparse high-dimensional linear regressionAnnals of Statistics (Ann. Stat.), 2020
I. Silin
Jianqing Fan
239
8
0
24 Jul 2020
Decentralised Learning with Random Features and Distributed Gradient
  Descent
Decentralised Learning with Random Features and Distributed Gradient Descent
Dominic Richards
Patrick Rebeschini
Lorenzo Rosasco
308
20
0
01 Jul 2020
Optimal Rates of Distributed Regression with Imperfect Kernels
Optimal Rates of Distributed Regression with Imperfect Kernels
Hongwei Sun
Qiang Wu
242
15
0
30 Jun 2020
On the Optimal Weighted $\ell_2$ Regularization in Overparameterized
  Linear Regression
On the Optimal Weighted ℓ2\ell_2ℓ2​ Regularization in Overparameterized Linear RegressionNeural Information Processing Systems (NeurIPS), 2020
Denny Wu
Ji Xu
522
144
0
10 Jun 2020
Optimal Regularization Can Mitigate Double Descent
Optimal Regularization Can Mitigate Double DescentInternational Conference on Learning Representations (ICLR), 2020
Preetum Nakkiran
Prayaag Venkat
Sham Kakade
Tengyu Ma
448
148
0
04 Mar 2020
Ridge Regression: Structure, Cross-Validation, and Sketching
Ridge Regression: Structure, Cross-Validation, and SketchingInternational Conference on Learning Representations (ICLR), 2019
Sifan Liu
Guang Cheng
CML
497
52
0
06 Oct 2019
Distributed linear regression by averaging
Distributed linear regression by averaging
Guang Cheng
Yueqi Sheng
FedML
477
75
0
30 Sep 2018
1
Page 1 of 1