ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1506.02617
  4. Cited By
Path-SGD: Path-Normalized Optimization in Deep Neural Networks

Path-SGD: Path-Normalized Optimization in Deep Neural Networks

Neural Information Processing Systems (NeurIPS), 2015
8 June 2015
Behnam Neyshabur
Ruslan Salakhutdinov
Nathan Srebro
    ODL
ArXiv (abs)PDFHTML

Papers citing "Path-SGD: Path-Normalized Optimization in Deep Neural Networks"

50 / 195 papers shown
On regularization of gradient descent, layer imbalance and flat minima
On regularization of gradient descent, layer imbalance and flat minima
Boris Ginsburg
36
2
0
18 Jul 2020
RIFLE: Backpropagation in Depth for Deep Transfer Learning through
  Re-Initializing the Fully-connected LayEr
RIFLE: Backpropagation in Depth for Deep Transfer Learning through Re-Initializing the Fully-connected LayEr
Xingjian Li
Haoyi Xiong
Haozhe An
Chengzhong Xu
Dejing Dou
ODL
127
43
0
07 Jul 2020
Using Human Psychophysics to Evaluate Generalization in Scene Text
  Recognition Models
Using Human Psychophysics to Evaluate Generalization in Scene Text Recognition Models
Sahar Siddiqui
E. Sizikova
Gemma Roig
N. Majaj
D. Pelli
87
1
0
30 Jun 2020
Learning compositional functions via multiplicative weight updates
Learning compositional functions via multiplicative weight updates
Jeremy Bernstein
Jiawei Zhao
M. Meister
Xuan Li
Anima Anandkumar
Yisong Yue
246
33
0
25 Jun 2020
Shape Matters: Understanding the Implicit Bias of the Noise Covariance
Shape Matters: Understanding the Implicit Bias of the Noise Covariance
Jeff Z. HaoChen
Colin Wei
Jason D. Lee
Tengyu Ma
615
109
0
15 Jun 2020
FLeet: Online Federated Learning via Staleness Awareness and Performance
  Prediction
FLeet: Online Federated Learning via Staleness Awareness and Performance PredictionInternational Middleware Conference (Middleware), 2020
Georgios Damaskinos
R. Guerraoui
Anne-Marie Kermarrec
Vlad Nitu
Rhicheek Patra
Francois Taiani
256
59
0
12 Jun 2020
Tangent Space Sensitivity and Distribution of Linear Regions in ReLU
  Networks
Tangent Space Sensitivity and Distribution of Linear Regions in ReLU Networks
Balint Daroczy
AAML
100
0
0
11 Jun 2020
Neural Path Features and Neural Path Kernel : Understanding the role of
  gates in deep learning
Neural Path Features and Neural Path Kernel : Understanding the role of gates in deep learningNeural Information Processing Systems (NeurIPS), 2020
Chandrashekar Lakshminarayanan
Amit Singh
AI4CE
183
11
0
11 Jun 2020
Banach Space Representer Theorems for Neural Networks and Ridge Splines
Banach Space Representer Theorems for Neural Networks and Ridge Splines
Rahul Parhi
Robert D. Nowak
226
7
0
10 Jun 2020
Pruning neural networks without any data by iteratively conserving
  synaptic flow
Pruning neural networks without any data by iteratively conserving synaptic flow
Hidenori Tanaka
D. Kunin
Daniel L. K. Yamins
Surya Ganguli
559
769
0
09 Jun 2020
Statistical Guarantees for Regularized Neural Networks
Statistical Guarantees for Regularized Neural NetworksNeural Networks (NN), 2020
Mahsa Taheri
Fang Xie
Johannes Lederer
278
41
0
30 May 2020
Scaling-up Distributed Processing of Data Streams for Machine Learning
Scaling-up Distributed Processing of Data Streams for Machine Learning
M. Nokleby
Haroon Raja
W. Bajwa
232
18
0
18 May 2020
Dropout: Explicit Forms and Capacity Control
Dropout: Explicit Forms and Capacity ControlInternational Conference on Machine Learning (ICML), 2020
R. Arora
Peter L. Bartlett
Poorya Mianjy
Nathan Srebro
213
41
0
06 Mar 2020
On the distance between two neural networks and the stability of
  learning
On the distance between two neural networks and the stability of learningNeural Information Processing Systems (NeurIPS), 2020
Jeremy Bernstein
Arash Vahdat
Yisong Yue
Xuan Li
ODL
507
70
0
09 Feb 2020
Understanding Generalization in Deep Learning via Tensor Methods
Understanding Generalization in Deep Learning via Tensor MethodsInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2020
Jingling Li
Yanchao Sun
Jiahao Su
Taiji Suzuki
Furong Huang
431
29
0
14 Jan 2020
Relative Flatness and Generalization
Relative Flatness and GeneralizationNeural Information Processing Systems (NeurIPS), 2020
Henning Petzka
Michael Kamp
Linara Adilova
C. Sminchisescu
Mario Boley
375
92
0
03 Jan 2020
Optimization for deep learning: theory and algorithms
Optimization for deep learning: theory and algorithms
Tian Ding
ODL
346
179
0
19 Dec 2019
A priori generalization error for two-layer ReLU neural network through minimum norm solution
Zhi-Qin John Xu
Jiwei Zhang
Yaoyu Zhang
Chengchao Zhao
MLT
184
1
0
06 Dec 2019
Fantastic Generalization Measures and Where to Find Them
Fantastic Generalization Measures and Where to Find ThemInternational Conference on Learning Representations (ICLR), 2019
Yiding Jiang
Behnam Neyshabur
H. Mobahi
Dilip Krishnan
Samy Bengio
AI4CE
462
673
0
04 Dec 2019
Information-Theoretic Local Minima Characterization and Regularization
Information-Theoretic Local Minima Characterization and RegularizationInternational Conference on Machine Learning (ICML), 2019
Zhiwei Jia
Hao Su
249
22
0
19 Nov 2019
On Generalization Bounds of a Family of Recurrent Neural Networks
On Generalization Bounds of a Family of Recurrent Neural NetworksInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2018
Minshuo Chen
Xingguo Li
T. Zhao
251
78
0
28 Oct 2019
Interpreting Basis Path Set in Neural Networks
Interpreting Basis Path Set in Neural Networks
Juanping Zhu
Qi Meng
Wei Chen
Zhi-Ming Ma
93
4
0
18 Oct 2019
Student Specialization in Deep ReLU Networks With Finite Width and Input
  Dimension
Student Specialization in Deep ReLU Networks With Finite Width and Input Dimension
Yuandong Tian
MLT
218
8
0
30 Sep 2019
Quantum Natural Gradient
Quantum Natural GradientQuantum (Quantum), 2019
J. Stokes
J. Izaac
N. Killoran
Giuseppe Carleo
241
479
0
04 Sep 2019
Gradient Descent Maximizes the Margin of Homogeneous Neural Networks
Gradient Descent Maximizes the Margin of Homogeneous Neural NetworksInternational Conference on Learning Representations (ICLR), 2019
Kaifeng Lyu
Jian Li
540
373
0
13 Jun 2019
The Implicit Bias of AdaGrad on Separable Data
The Implicit Bias of AdaGrad on Separable DataNeural Information Processing Systems (NeurIPS), 2019
Qian Qian
Xiaoyuan Qian
135
24
0
09 Jun 2019
Inductive Bias of Gradient Descent based Adversarial Training on
  Separable Data
Inductive Bias of Gradient Descent based Adversarial Training on Separable Data
Yan Li
Ethan X. Fang
Huan Xu
T. Zhao
273
18
0
07 Jun 2019
On Dropout and Nuclear Norm Regularization
On Dropout and Nuclear Norm RegularizationInternational Conference on Machine Learning (ICML), 2019
Poorya Mianjy
R. Arora
264
24
0
28 May 2019
Quantifying the generalization error in deep learning in terms of data
  distribution and neural network smoothness
Quantifying the generalization error in deep learning in terms of data distribution and neural network smoothnessNeural Networks (NN), 2019
Pengzhan Jin
Lu Lu
Yifa Tang
George Karniadakis
351
65
0
27 May 2019
Exploring Structural Sparsity of Deep Networks via Inverse Scale Spaces
Exploring Structural Sparsity of Deep Networks via Inverse Scale SpacesIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2019
Yanwei Fu
Chen Liu
Donghao Li
Zuyuan Zhong
Xinwei Sun
Jinshan Zeng
Xingtai Lv
261
14
0
23 May 2019
Lexicographic and Depth-Sensitive Margins in Homogeneous and
  Non-Homogeneous Deep Models
Lexicographic and Depth-Sensitive Margins in Homogeneous and Non-Homogeneous Deep ModelsInternational Conference on Machine Learning (ICML), 2019
Mor Shpigel Nacson
Suriya Gunasekar
Jason D. Lee
Nathan Srebro
Daniel Soudry
204
96
0
17 May 2019
Implicit Regularization of Discrete Gradient Dynamics in Linear Neural
  Networks
Implicit Regularization of Discrete Gradient Dynamics in Linear Neural NetworksNeural Information Processing Systems (NeurIPS), 2019
Gauthier Gidel
Francis R. Bach
Damien Scieur
AI4CE
204
170
0
30 Apr 2019
Iterative Normalization: Beyond Standardization towards Efficient
  Whitening
Iterative Normalization: Beyond Standardization towards Efficient Whitening
Lei Huang
Yi Zhou
Fan Zhu
Li Liu
Ling Shao
196
177
0
06 Apr 2019
Positively Scale-Invariant Flatness of ReLU Neural Networks
Positively Scale-Invariant Flatness of ReLU Neural Networks
Mingyang Yi
Qi Meng
Wei-neng Chen
Zhi-Ming Ma
Tie-Yan Liu
174
19
0
06 Mar 2019
A Priori Estimates of the Population Risk for Residual Networks
A Priori Estimates of the Population Risk for Residual Networks
E. Weinan
Chao Ma
Qingcan Wang
UQCV
222
62
0
06 Mar 2019
Equi-normalization of Neural Networks
Equi-normalization of Neural NetworksInternational Conference on Learning Representations (ICLR), 2019
Pierre Stock
Benjamin Graham
Rémi Gribonval
Edouard Grave
ODL
144
19
0
27 Feb 2019
A Scale Invariant Flatness Measure for Deep Network Minima
A Scale Invariant Flatness Measure for Deep Network Minima
Akshay Rangamani
Nam H. Nguyen
Abhishek Kumar
Dzung Phan
Sang H. Chin
T. Tran
ODL
163
32
0
06 Feb 2019
Are All Layers Created Equal?
Are All Layers Created Equal?
Chiyuan Zhang
Samy Bengio
Y. Singer
337
158
0
06 Feb 2019
Trajectory Normalized Gradients for Distributed Optimization
Trajectory Normalized Gradients for Distributed Optimization
Jianqiao Wangni
Ke Li
Jianbo Shi
Jitendra Malik
132
2
0
24 Jan 2019
A Theoretical Analysis of Deep Q-Learning
A Theoretical Analysis of Deep Q-Learning
Jianqing Fan
Zhuoran Yang
Yuchen Xie
Zhaoran Wang
604
711
0
01 Jan 2019
A Differential Topological View of Challenges in Learning with
  Feedforward Neural Networks
A Differential Topological View of Challenges in Learning with Feedforward Neural Networks
Hao Shen
AAMLAI4CE
148
6
0
26 Nov 2018
Deep Frank-Wolfe For Neural Network Optimization
Deep Frank-Wolfe For Neural Network OptimizationInternational Conference on Learning Representations (ICLR), 2018
Leonard Berrada
Andrew Zisserman
M. P. Kumar
ODL
204
41
0
19 Nov 2018
A Bayesian Perspective of Convolutional Neural Networks through a
  Deconvolutional Generative Model
A Bayesian Perspective of Convolutional Neural Networks through a Deconvolutional Generative Model
Yujia Wang
Nhat Ho
David J. Miller
Anima Anandkumar
Sai Li
Richard G. Baraniuk
BDLGAN
250
8
0
01 Nov 2018
The loss surface of deep linear networks viewed through the algebraic
  geometry lens
The loss surface of deep linear networks viewed through the algebraic geometry lens
D. Mehta
Tianran Chen
Tingting Tang
J. Hauenstein
ODL
233
35
0
17 Oct 2018
Regularization Matters: Generalization and Optimization of Neural Nets
  v.s. their Induced Kernel
Regularization Matters: Generalization and Optimization of Neural Nets v.s. their Induced Kernel
Colin Wei
Jason D. Lee
Qiang Liu
Tengyu Ma
735
263
0
12 Oct 2018
Capacity Control of ReLU Neural Networks by Basis-path Norm
Capacity Control of ReLU Neural Networks by Basis-path Norm
Shuxin Zheng
Qi Meng
Huishuai Zhang
Wei-neng Chen
Nenghai Yu
Tie-Yan Liu
159
22
0
19 Sep 2018
Approximation and Estimation for High-Dimensional Deep Learning Networks
Approximation and Estimation for High-Dimensional Deep Learning Networks
Andrew R. Barron
Jason M. Klusowski
222
60
0
10 Sep 2018
Deep Neural Networks with Multi-Branch Architectures Are Less Non-Convex
Deep Neural Networks with Multi-Branch Architectures Are Less Non-Convex
Hongyang R. Zhang
Junru Shao
Ruslan Salakhutdinov
285
14
0
06 Jun 2018
Algorithmic Regularization in Learning Deep Homogeneous Models: Layers
  are Automatically Balanced
Algorithmic Regularization in Learning Deep Homogeneous Models: Layers are Automatically Balanced
S. Du
Wei Hu
Jason D. Lee
MLT
465
264
0
04 Jun 2018
Implicit Bias of Gradient Descent on Linear Convolutional Networks
Implicit Bias of Gradient Descent on Linear Convolutional Networks
Suriya Gunasekar
Jason D. Lee
Daniel Soudry
Nathan Srebro
MDE
468
444
0
01 Jun 2018
Previous
1234
Next
Page 3 of 4