Parallelizing Stochastic Gradient Descent for Least Squares Regression:
  mini-batching, averaging, and model misspecification
v1v2v3v4 (latest)

Parallelizing Stochastic Gradient Descent for Least Squares Regression: mini-batching, averaging, and model misspecification

    MoMe

Papers citing "Parallelizing Stochastic Gradient Descent for Least Squares Regression: mini-batching, averaging, and model misspecification"

21 / 21 papers shown
Title
Anytime Tail Averaging
Anytime Tail Averaging
Nicolas Le Roux
87
5
0
13 Feb 2019
Online to Offline Conversions, Universality and Adaptive Minibatch Sizes
Online to Offline Conversions, Universality and Adaptive Minibatch SizesNeural Information Processing Systems (NeurIPS), 2017
179
64
0
30 May 2017
Stochastic Composite Least-Squares Regression with convergence rate
  O(1/n)
Stochastic Composite Least-Squares Regression with convergence rate O(1/n)Annual Conference Computational Learning Theory (COLT), 2017
120
28
0
21 Feb 2017

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.