Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2004.06977
Cited By
On Learning Rates and Schrödinger Operators
Journal of machine learning research (JMLR), 2020
15 April 2020
Bin Shi
Weijie J. Su
Sai Li
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"On Learning Rates and Schrödinger Operators"
36 / 36 papers shown
Statistical Guarantees for High-Dimensional Stochastic Gradient Descent
Jiaqi Li
Zhipeng Lou
Johannes Schmidt-Hieber
Wei Biao Wu
177
0
0
13 Oct 2025
A Spin Glass Characterization of Neural Networks
Jun Li
171
1
0
10 Aug 2025
Quantum Optimization via Gradient-Based Hamiltonian Descent
Jiaqi Leng
Bin Shi
287
4
0
20 May 2025
FOCUS: First Order Concentrated Updating Scheme
Yizhou Liu
Ziming Liu
Jeff Gore
ODL
427
4
0
21 Jan 2025
Where Do Large Learning Rates Lead Us?
Neural Information Processing Systems (NeurIPS), 2024
Ildus Sadrtdinov
M. Kodryan
Eduard Pokonechny
E. Lobacheva
Dmitry Vetrov
AI4CE
378
6
0
29 Oct 2024
Stepping on the Edge: Curvature Aware Learning Rate Tuners
Vincent Roulet
Atish Agarwala
Jean-Bastien Grill
Grzegorz Swirszcz
Mathieu Blondel
Fabian Pedregosa
459
6
0
08 Jul 2024
Complex fractal trainability boundary can arise from trivial non-convexity
Yizhou Liu
203
1
0
20 Jun 2024
Primitive Agentic First-Order Optimization
International Conference on Control, Decision and Information Technologies (CoDIT), 2024
R. Sala
260
0
0
07 Jun 2024
A General Continuous-Time Formulation of Stochastic ADMM and Its Variants
Chris Junchi Li
283
1
0
22 Apr 2024
A simple lower bound for the complexity of estimating partition functions on a quantum computer
Zherui Chen
Giacomo Nannicini
266
1
0
03 Apr 2024
Arbitrary Polynomial Separations in Trainable Quantum Machine Learning
Eric R. Anschuetz
Xun Gao
200
10
0
13 Feb 2024
Quantum Langevin Dynamics for Optimization
Communications in Mathematical Physics (CMP), 2023
Zherui Chen
Yuchen Lu
Hao Wang
Yizhou Liu
Tongyang Li
AI4CE
631
22
0
27 Nov 2023
A quantum-classical performance separation in nonconvex optimization
Jiaqi Leng
Yufan Zheng
Xiaodi Wu
284
11
0
01 Nov 2023
Global Convergence of SGD For Logistic Loss on Two Layer Neural Nets
Pulkit Gopalani
Samyak Jha
Anirbit Mukherjee
335
3
0
17 Sep 2023
On Underdamped Nesterov's Acceleration
Shu Chen
Bin Shi
Ya-xiang Yuan
299
5
0
28 Apr 2023
Training a Two Layer ReLU Network Analytically
Italian National Conference on Sensors (INS), 2023
Adrian Barbu
319
9
0
06 Apr 2023
Learning Rate Schedules in the Presence of Distribution Shift
International Conference on Machine Learning (ICML), 2023
Matthew Fahrbach
Adel Javanmard
Vahab Mirrokni
Pratik Worah
298
11
0
27 Mar 2023
Global Convergence of SGD On Two Layer Neural Nets
Information and Inference A Journal of the IMA (JIII), 2022
Pulkit Gopalani
Anirbit Mukherjee
268
9
0
20 Oct 2022
On Quantum Speedups for Nonconvex Optimization via Quantum Tunneling Walks
Quantum (Quantum), 2022
Yizhou Liu
Weijie J. Su
Tongyang Li
357
25
0
29 Sep 2022
Gradient Norm Minimization of Nesterov Acceleration:
o
(
1
/
k
3
)
o(1/k^3)
o
(
1/
k
3
)
Shu Chen
Bin Shi
Ya-xiang Yuan
292
20
0
19 Sep 2022
On Uniform Boundedness Properties of SGD and its Momentum Variants
Xiaoyu Wang
M. Johansson
242
3
0
25 Jan 2022
Imitating Deep Learning Dynamics via Locally Elastic Stochastic Differential Equations
Neural Information Processing Systems (NeurIPS), 2021
Jiayao Zhang
Hua Wang
Weijie J. Su
326
9
0
11 Oct 2021
On the Hyperparameters in Stochastic Gradient Descent with Momentum
Journal of machine learning research (JMLR), 2021
Bin Shi
309
20
0
09 Aug 2021
Revisiting the Effects of Stochasticity for Hamiltonian Samplers
International Conference on Machine Learning (ICML), 2021
Giulio Franzese
Dimitrios Milios
Maurizio Filippone
Pietro Michiardi
267
3
0
30 Jun 2021
Bandwidth-based Step-Sizes for Non-Convex Stochastic Optimization
Xiaoyu Wang
M. Johansson
293
2
0
05 Jun 2021
On the Validity of Modeling SGD with Stochastic Differential Equations (SDEs)
Neural Information Processing Systems (NeurIPS), 2021
Zhiyuan Li
Sadhika Malladi
Sanjeev Arora
305
100
0
24 Feb 2021
On the Convergence of Step Decay Step-Size for Stochastic Optimization
Neural Information Processing Systems (NeurIPS), 2021
Xiaoyu Wang
Sindri Magnússon
M. Johansson
278
31
0
18 Feb 2021
Exploring Deep Neural Networks via Layer-Peeled Model: Minority Collapse in Imbalanced Training
Proceedings of the National Academy of Sciences of the United States of America (PNAS), 2021
Cong Fang
Hangfeng He
Qi Long
Weijie J. Su
FAtt
519
217
0
29 Jan 2021
Boundary Conditions for Linear Exit Time Gradient Trajectories Around Saddle Points: Analysis and Algorithm
IEEE Transactions on Information Theory (IEEE Trans. Inf. Theory), 2021
Rishabh Dixit
Mert Gurbuzbalaban
W. Bajwa
280
2
0
07 Jan 2021
Loss Barcode: A Topological Measure of Escapability in Loss Landscapes
S. Barannikov
Daria Voronkova
I. Trofimov
Alexander Korotin
Grigorii Sotnikov
Evgeny Burnaev
Evgeny Burnaev
284
7
0
31 Dec 2020
Approximation of BV functions by neural networks: A regularity theory approach
Analysis and Applications (Anal. Appl.), 2020
B. Avelin
Vesa Julin
148
3
0
15 Dec 2020
Reconciling Modern Deep Learning with Traditional Optimization Analyses: The Intrinsic Learning Rate
Zhiyuan Li
Kaifeng Lyu
Sanjeev Arora
300
78
0
06 Oct 2020
Group Knowledge Transfer: Federated Learning of Large CNNs at the Edge
Chaoyang He
M. Annavaram
A. Avestimehr
FedML
386
24
0
28 Jul 2020
On stochastic mirror descent with interacting particles: convergence properties and variance reduction
Anastasia Borovykh
N. Kantas
P. Parpas
G. Pavliotis
246
13
0
15 Jul 2020
Exit Time Analysis for Approximations of Gradient Descent Trajectories Around Saddle Points
Rishabh Dixit
Mert Gurbuzbalaban
W. Bajwa
278
5
0
01 Jun 2020
Scaling-up Distributed Processing of Data Streams for Machine Learning
M. Nokleby
Haroon Raja
W. Bajwa
278
18
0
18 May 2020
1
Page 1 of 1