ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2004.06977
  4. Cited By
On Learning Rates and Schrödinger Operators

On Learning Rates and Schrödinger Operators

Journal of machine learning research (JMLR), 2020
15 April 2020
Bin Shi
Weijie J. Su
Sai Li
ArXiv (abs)PDFHTML

Papers citing "On Learning Rates and Schrödinger Operators"

36 / 36 papers shown
Statistical Guarantees for High-Dimensional Stochastic Gradient Descent
Statistical Guarantees for High-Dimensional Stochastic Gradient Descent
Jiaqi Li
Zhipeng Lou
Johannes Schmidt-Hieber
Wei Biao Wu
177
0
0
13 Oct 2025
A Spin Glass Characterization of Neural Networks
A Spin Glass Characterization of Neural Networks
Jun Li
171
1
0
10 Aug 2025
Quantum Optimization via Gradient-Based Hamiltonian Descent
Quantum Optimization via Gradient-Based Hamiltonian Descent
Jiaqi Leng
Bin Shi
287
4
0
20 May 2025
FOCUS: First Order Concentrated Updating Scheme
FOCUS: First Order Concentrated Updating Scheme
Yizhou Liu
Ziming Liu
Jeff Gore
ODL
427
4
0
21 Jan 2025
Where Do Large Learning Rates Lead Us?
Where Do Large Learning Rates Lead Us?Neural Information Processing Systems (NeurIPS), 2024
Ildus Sadrtdinov
M. Kodryan
Eduard Pokonechny
E. Lobacheva
Dmitry Vetrov
AI4CE
378
6
0
29 Oct 2024
Stepping on the Edge: Curvature Aware Learning Rate Tuners
Stepping on the Edge: Curvature Aware Learning Rate Tuners
Vincent Roulet
Atish Agarwala
Jean-Bastien Grill
Grzegorz Swirszcz
Mathieu Blondel
Fabian Pedregosa
459
6
0
08 Jul 2024
Complex fractal trainability boundary can arise from trivial
  non-convexity
Complex fractal trainability boundary can arise from trivial non-convexity
Yizhou Liu
203
1
0
20 Jun 2024
Primitive Agentic First-Order Optimization
Primitive Agentic First-Order OptimizationInternational Conference on Control, Decision and Information Technologies (CoDIT), 2024
R. Sala
260
0
0
07 Jun 2024
A General Continuous-Time Formulation of Stochastic ADMM and Its
  Variants
A General Continuous-Time Formulation of Stochastic ADMM and Its Variants
Chris Junchi Li
283
1
0
22 Apr 2024
A simple lower bound for the complexity of estimating partition
  functions on a quantum computer
A simple lower bound for the complexity of estimating partition functions on a quantum computer
Zherui Chen
Giacomo Nannicini
266
1
0
03 Apr 2024
Arbitrary Polynomial Separations in Trainable Quantum Machine Learning
Arbitrary Polynomial Separations in Trainable Quantum Machine Learning
Eric R. Anschuetz
Xun Gao
200
10
0
13 Feb 2024
Quantum Langevin Dynamics for Optimization
Quantum Langevin Dynamics for OptimizationCommunications in Mathematical Physics (CMP), 2023
Zherui Chen
Yuchen Lu
Hao Wang
Yizhou Liu
Tongyang Li
AI4CE
631
22
0
27 Nov 2023
A quantum-classical performance separation in nonconvex optimization
A quantum-classical performance separation in nonconvex optimization
Jiaqi Leng
Yufan Zheng
Xiaodi Wu
284
11
0
01 Nov 2023
Global Convergence of SGD For Logistic Loss on Two Layer Neural Nets
Global Convergence of SGD For Logistic Loss on Two Layer Neural Nets
Pulkit Gopalani
Samyak Jha
Anirbit Mukherjee
335
3
0
17 Sep 2023
On Underdamped Nesterov's Acceleration
On Underdamped Nesterov's Acceleration
Shu Chen
Bin Shi
Ya-xiang Yuan
299
5
0
28 Apr 2023
Training a Two Layer ReLU Network Analytically
Training a Two Layer ReLU Network AnalyticallyItalian National Conference on Sensors (INS), 2023
Adrian Barbu
319
9
0
06 Apr 2023
Learning Rate Schedules in the Presence of Distribution Shift
Learning Rate Schedules in the Presence of Distribution ShiftInternational Conference on Machine Learning (ICML), 2023
Matthew Fahrbach
Adel Javanmard
Vahab Mirrokni
Pratik Worah
298
11
0
27 Mar 2023
Global Convergence of SGD On Two Layer Neural Nets
Global Convergence of SGD On Two Layer Neural NetsInformation and Inference A Journal of the IMA (JIII), 2022
Pulkit Gopalani
Anirbit Mukherjee
268
9
0
20 Oct 2022
On Quantum Speedups for Nonconvex Optimization via Quantum Tunneling
  Walks
On Quantum Speedups for Nonconvex Optimization via Quantum Tunneling WalksQuantum (Quantum), 2022
Yizhou Liu
Weijie J. Su
Tongyang Li
357
25
0
29 Sep 2022
Gradient Norm Minimization of Nesterov Acceleration: $o(1/k^3)$
Gradient Norm Minimization of Nesterov Acceleration: o(1/k3)o(1/k^3)o(1/k3)
Shu Chen
Bin Shi
Ya-xiang Yuan
292
20
0
19 Sep 2022
On Uniform Boundedness Properties of SGD and its Momentum Variants
On Uniform Boundedness Properties of SGD and its Momentum Variants
Xiaoyu Wang
M. Johansson
242
3
0
25 Jan 2022
Imitating Deep Learning Dynamics via Locally Elastic Stochastic
  Differential Equations
Imitating Deep Learning Dynamics via Locally Elastic Stochastic Differential EquationsNeural Information Processing Systems (NeurIPS), 2021
Jiayao Zhang
Hua Wang
Weijie J. Su
326
9
0
11 Oct 2021
On the Hyperparameters in Stochastic Gradient Descent with Momentum
On the Hyperparameters in Stochastic Gradient Descent with MomentumJournal of machine learning research (JMLR), 2021
Bin Shi
309
20
0
09 Aug 2021
Revisiting the Effects of Stochasticity for Hamiltonian Samplers
Revisiting the Effects of Stochasticity for Hamiltonian SamplersInternational Conference on Machine Learning (ICML), 2021
Giulio Franzese
Dimitrios Milios
Maurizio Filippone
Pietro Michiardi
267
3
0
30 Jun 2021
Bandwidth-based Step-Sizes for Non-Convex Stochastic Optimization
Bandwidth-based Step-Sizes for Non-Convex Stochastic Optimization
Xiaoyu Wang
M. Johansson
293
2
0
05 Jun 2021
On the Validity of Modeling SGD with Stochastic Differential Equations
  (SDEs)
On the Validity of Modeling SGD with Stochastic Differential Equations (SDEs)Neural Information Processing Systems (NeurIPS), 2021
Zhiyuan Li
Sadhika Malladi
Sanjeev Arora
305
100
0
24 Feb 2021
On the Convergence of Step Decay Step-Size for Stochastic Optimization
On the Convergence of Step Decay Step-Size for Stochastic OptimizationNeural Information Processing Systems (NeurIPS), 2021
Xiaoyu Wang
Sindri Magnússon
M. Johansson
278
31
0
18 Feb 2021
Exploring Deep Neural Networks via Layer-Peeled Model: Minority Collapse
  in Imbalanced Training
Exploring Deep Neural Networks via Layer-Peeled Model: Minority Collapse in Imbalanced TrainingProceedings of the National Academy of Sciences of the United States of America (PNAS), 2021
Cong Fang
Hangfeng He
Qi Long
Weijie J. Su
FAtt
519
217
0
29 Jan 2021
Boundary Conditions for Linear Exit Time Gradient Trajectories Around
  Saddle Points: Analysis and Algorithm
Boundary Conditions for Linear Exit Time Gradient Trajectories Around Saddle Points: Analysis and AlgorithmIEEE Transactions on Information Theory (IEEE Trans. Inf. Theory), 2021
Rishabh Dixit
Mert Gurbuzbalaban
W. Bajwa
280
2
0
07 Jan 2021
Loss Barcode: A Topological Measure of Escapability in Loss Landscapes
Loss Barcode: A Topological Measure of Escapability in Loss Landscapes
S. Barannikov
Daria Voronkova
I. Trofimov
Alexander Korotin
Grigorii Sotnikov
Evgeny Burnaev
Evgeny Burnaev
284
7
0
31 Dec 2020
Approximation of BV functions by neural networks: A regularity theory
  approach
Approximation of BV functions by neural networks: A regularity theory approachAnalysis and Applications (Anal. Appl.), 2020
B. Avelin
Vesa Julin
148
3
0
15 Dec 2020
Reconciling Modern Deep Learning with Traditional Optimization Analyses:
  The Intrinsic Learning Rate
Reconciling Modern Deep Learning with Traditional Optimization Analyses: The Intrinsic Learning Rate
Zhiyuan Li
Kaifeng Lyu
Sanjeev Arora
300
78
0
06 Oct 2020
Group Knowledge Transfer: Federated Learning of Large CNNs at the Edge
Group Knowledge Transfer: Federated Learning of Large CNNs at the Edge
Chaoyang He
M. Annavaram
A. Avestimehr
FedML
386
24
0
28 Jul 2020
On stochastic mirror descent with interacting particles: convergence
  properties and variance reduction
On stochastic mirror descent with interacting particles: convergence properties and variance reduction
Anastasia Borovykh
N. Kantas
P. Parpas
G. Pavliotis
246
13
0
15 Jul 2020
Exit Time Analysis for Approximations of Gradient Descent Trajectories
  Around Saddle Points
Exit Time Analysis for Approximations of Gradient Descent Trajectories Around Saddle Points
Rishabh Dixit
Mert Gurbuzbalaban
W. Bajwa
278
5
0
01 Jun 2020
Scaling-up Distributed Processing of Data Streams for Machine Learning
Scaling-up Distributed Processing of Data Streams for Machine Learning
M. Nokleby
Haroon Raja
W. Bajwa
278
18
0
18 May 2020
1
Page 1 of 1