ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1802.06384
  4. Cited By
Spurious Valleys in Two-layer Neural Network Optimization Landscapes
v1v2v3v4 (latest)

Spurious Valleys in Two-layer Neural Network Optimization Landscapes

18 February 2018
Luca Venturi
Afonso S. Bandeira
Joan Bruna
ArXiv (abs)PDFHTML

Papers citing "Spurious Valleys in Two-layer Neural Network Optimization Landscapes"

50 / 51 papers shown
A topological description of loss surfaces based on Betti Numbers
A topological description of loss surfaces based on Betti NumbersNeural Networks (NN), 2024
Maria Sofia Bucarelli
Giuseppe Alessio D’Inverno
Monica Bianchini
F. Scarselli
Fabrizio Silvestri
181
5
0
08 Jan 2024
Minimum norm interpolation by perceptra: Explicit regularization and
  implicit bias
Minimum norm interpolation by perceptra: Explicit regularization and implicit biasNeural Information Processing Systems (NeurIPS), 2023
Jiyoung Park
Ian Pelakh
Stephan Wojtowytsch
251
2
0
10 Nov 2023
A qualitative difference between gradient flows of convex functions in
  finite- and infinite-dimensional Hilbert spaces
A qualitative difference between gradient flows of convex functions in finite- and infinite-dimensional Hilbert spaces
Jonathan W. Siegel
Stephan Wojtowytsch
275
5
0
26 Oct 2023
NTK-SAP: Improving neural network pruning by aligning training dynamics
NTK-SAP: Improving neural network pruning by aligning training dynamicsInternational Conference on Learning Representations (ICLR), 2023
Yite Wang
Dawei Li
Tian Ding
329
34
0
06 Apr 2023
When Expressivity Meets Trainability: Fewer than $n$ Neurons Can Work
When Expressivity Meets Trainability: Fewer than nnn Neurons Can WorkNeural Information Processing Systems (NeurIPS), 2022
Jiawei Zhang
Yushun Zhang
Mingyi Hong
Tian Ding
Jianfeng Yao
366
11
0
21 Oct 2022
Plateau in Monotonic Linear Interpolation -- A "Biased" View of Loss
  Landscape for Deep Networks
Plateau in Monotonic Linear Interpolation -- A "Biased" View of Loss Landscape for Deep NetworksInternational Conference on Learning Representations (ICLR), 2022
Xiang Wang
Annie Wang
Mo Zhou
Rong Ge
MoMe
556
11
0
03 Oct 2022
Deep Architecture Connectivity Matters for Its Convergence: A
  Fine-Grained Analysis
Deep Architecture Connectivity Matters for Its Convergence: A Fine-Grained AnalysisNeural Information Processing Systems (NeurIPS), 2022
Wuyang Chen
Wei-Ping Huang
Xinyu Gong
Boris Hanin
Zinan Lin
319
9
0
11 May 2022
Deep learning, stochastic gradient descent and diffusion maps
Deep learning, stochastic gradient descent and diffusion mapsJournal of Computational Mathematics and Data Science (JCMDS), 2022
Carmina Fjellström
Kaj Nyström
DiffM
329
22
0
04 Apr 2022
Global Convergence Analysis of Deep Linear Networks with A One-neuron
  Layer
Global Convergence Analysis of Deep Linear Networks with A One-neuron Layer
Kun Chen
Dachao Lin
Zhihua Zhang
234
1
0
08 Jan 2022
Complexity from Adaptive-Symmetries Breaking: Global Minima in the
  Statistical Mechanics of Deep Neural Networks
Complexity from Adaptive-Symmetries Breaking: Global Minima in the Statistical Mechanics of Deep Neural Networks
Shaun Li
AI4CE
256
1
0
03 Jan 2022
What Happens after SGD Reaches Zero Loss? --A Mathematical Framework
What Happens after SGD Reaches Zero Loss? --A Mathematical Framework
Zhiyuan Li
Tianhao Wang
Sanjeev Arora
MLT
472
119
0
13 Oct 2021
Exponentially Many Local Minima in Quantum Neural Networks
Exponentially Many Local Minima in Quantum Neural Networks
Xuchen You
Xiaodi Wu
366
64
0
06 Oct 2021
Analyzing Monotonic Linear Interpolation in Neural Network Loss
  Landscapes
Analyzing Monotonic Linear Interpolation in Neural Network Loss Landscapes
James Lucas
Juhan Bae
Michael Ruogu Zhang
Stanislav Fort
R. Zemel
Roger C. Grosse
MoMe
552
32
0
22 Apr 2021
Spurious Local Minima Are Common for Deep Neural Networks with Piecewise
  Linear Activations
Spurious Local Minima Are Common for Deep Neural Networks with Piecewise Linear ActivationsIEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2021
Bo Liu
166
12
0
25 Feb 2021
Noisy Gradient Descent Converges to Flat Minima for Nonconvex Matrix
  Factorization
Noisy Gradient Descent Converges to Flat Minima for Nonconvex Matrix FactorizationInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2021
Tianyi Liu
Yan Li
S. Wei
Enlu Zhou
T. Zhao
238
19
0
24 Feb 2021
A Note on Connectivity of Sublevel Sets in Deep Learning
A Note on Connectivity of Sublevel Sets in Deep Learning
Quynh N. Nguyen
MLT
319
9
0
21 Jan 2021
A Convergence Theory Towards Practical Over-parameterized Deep Neural
  Networks
A Convergence Theory Towards Practical Over-parameterized Deep Neural Networks
Asaf Noy
Yi Tian Xu
Y. Aflalo
Lihi Zelnik-Manor
Rong Jin
285
3
0
12 Jan 2021
Towards a Better Global Loss Landscape of GANs
Towards a Better Global Loss Landscape of GANs
Tian Ding
Tiantian Fang
Alex Schwing
GAN
292
36
0
10 Nov 2020
DessiLBI: Exploring Structural Sparsity of Deep Networks via
  Differential Inclusion Paths
DessiLBI: Exploring Structural Sparsity of Deep Networks via Differential Inclusion Paths
Yanwei Fu
Chen Liu
Donghao Li
Xinwei Sun
Jinshan Zeng
Xingtai Lv
228
9
0
04 Jul 2020
PDE constraints on smooth hierarchical functions computed by neural
  networks
PDE constraints on smooth hierarchical functions computed by neural networks
Khashayar Filom
Konrad Paul Kording
Roozbeh Farhoodi
156
0
0
18 May 2020
The critical locus of overparameterized neural networks
The critical locus of overparameterized neural networks
Y. Cooper
UQCV
265
11
0
08 May 2020
Compressive sensing with un-trained neural networks: Gradient descent
  finds the smoothest approximation
Compressive sensing with un-trained neural networks: Gradient descent finds the smoothest approximation
Reinhard Heckel
Mahdi Soltanolkotabi
200
91
0
07 May 2020
Some Geometrical and Topological Properties of DNNs' Decision Boundaries
Some Geometrical and Topological Properties of DNNs' Decision Boundaries
Bo Liu
Mengya Shen
AAML
280
3
0
07 Mar 2020
Learning the mapping $\mathbf{x}\mapsto \sum_{i=1}^d x_i^2$: the cost of
  finding the needle in a haystack
Learning the mapping x↦∑i=1dxi2\mathbf{x}\mapsto \sum_{i=1}^d x_i^2x↦∑i=1d​xi2​: the cost of finding the needle in a haystack
Jiefu Zhang
Leonardo Zepeda-Núnez
Xingtai Lv
Lin Lin
129
0
0
24 Feb 2020
Understanding Global Loss Landscape of One-hidden-layer ReLU Networks,
  Part 1: Theory
Understanding Global Loss Landscape of One-hidden-layer ReLU Networks, Part 1: Theory
Bo Liu
FAttMLT
431
1
0
12 Feb 2020
Landscape Connectivity and Dropout Stability of SGD Solutions for
  Over-parameterized Neural Networks
Landscape Connectivity and Dropout Stability of SGD Solutions for Over-parameterized Neural NetworksInternational Conference on Machine Learning (ICML), 2019
Aleksandr Shevchenko
Marco Mondelli
487
41
0
20 Dec 2019
Optimization for deep learning: theory and algorithms
Optimization for deep learning: theory and algorithms
Tian Ding
ODL
485
183
0
19 Dec 2019
Stationary Points of Shallow Neural Networks with Quadratic Activation
  Function
Stationary Points of Shallow Neural Networks with Quadratic Activation Function
D. Gamarnik
Eren C. Kizildag
Ilias Zadik
256
16
0
03 Dec 2019
Sub-Optimal Local Minima Exist for Neural Networks with Almost All
  Non-Linear Activations
Sub-Optimal Local Minima Exist for Neural Networks with Almost All Non-Linear Activations
Tian Ding
Dawei Li
Tian Ding
404
14
0
04 Nov 2019
Denoising and Regularization via Exploiting the Structural Bias of
  Convolutional Generators
Denoising and Regularization via Exploiting the Structural Bias of Convolutional GeneratorsInternational Conference on Learning Representations (ICLR), 2019
Reinhard Heckel
Mahdi Soltanolkotabi
DiffM
301
91
0
31 Oct 2019
Nearly Minimal Over-Parametrization of Shallow Neural Networks
Armin Eftekhari
Chaehwan Song
Volkan Cevher
226
1
0
09 Oct 2019
Pure and Spurious Critical Points: a Geometric Study of Linear Networks
Pure and Spurious Critical Points: a Geometric Study of Linear NetworksInternational Conference on Learning Representations (ICLR), 2019
Matthew Trager
Kathlén Kohn
Joan Bruna
266
41
0
03 Oct 2019
Generating Accurate Pseudo-labels in Semi-Supervised Learning and
  Avoiding Overconfident Predictions via Hermite Polynomial Activations
Generating Accurate Pseudo-labels in Semi-Supervised Learning and Avoiding Overconfident Predictions via Hermite Polynomial ActivationsComputer Vision and Pattern Recognition (CVPR), 2019
Vishnu Suresh Lokhande
Songwong Tasneeyapant
Abhay Venkatesh
Sathya Ravi
Vikas Singh
207
30
0
12 Sep 2019
Additive function approximation in the brain
Additive function approximation in the brain
K. Harris
262
15
0
05 Sep 2019
Gradient Dynamics of Shallow Univariate ReLU Networks
Gradient Dynamics of Shallow Univariate ReLU NetworksNeural Information Processing Systems (NeurIPS), 2019
Francis Williams
Matthew Trager
Claudio Silva
Daniele Panozzo
Denis Zorin
Joan Bruna
204
87
0
18 Jun 2019
Explaining Landscape Connectivity of Low-cost Solutions for Multilayer
  Nets
Explaining Landscape Connectivity of Low-cost Solutions for Multilayer NetsNeural Information Processing Systems (NeurIPS), 2019
Rohith Kuditipudi
Xiang Wang
Holden Lee
Yi Zhang
Zhiyuan Li
Wei Hu
Sanjeev Arora
Rong Ge
FAtt
541
102
0
14 Jun 2019
A mean-field limit for certain deep neural networks
A mean-field limit for certain deep neural networks
Dyego Araújo
R. Oliveira
Daniel Yukimura
AI4CE
380
71
0
01 Jun 2019
On the Expressive Power of Deep Polynomial Neural Networks
On the Expressive Power of Deep Polynomial Neural NetworksNeural Information Processing Systems (NeurIPS), 2019
Joe Kileel
Matthew Trager
Joan Bruna
250
102
0
29 May 2019
Exploring Structural Sparsity of Deep Networks via Inverse Scale Spaces
Exploring Structural Sparsity of Deep Networks via Inverse Scale SpacesIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2019
Yanwei Fu
Chen Liu
Donghao Li
Zuyuan Zhong
Xinwei Sun
Jinshan Zeng
Xingtai Lv
361
14
0
23 May 2019
The Landscape of the Planted Clique Problem: Dense subgraphs and the
  Overlap Gap Property
The Landscape of the Planted Clique Problem: Dense subgraphs and the Overlap Gap Property
D. Gamarnik
Ilias Zadik
217
42
0
15 Apr 2019
Gradient Descent with Early Stopping is Provably Robust to Label Noise
  for Overparameterized Neural Networks
Gradient Descent with Early Stopping is Provably Robust to Label Noise for Overparameterized Neural Networks
Mingchen Li
Mahdi Soltanolkotabi
Samet Oymak
NoLa
589
392
0
27 Mar 2019
Towards moderate overparameterization: global convergence guarantees for
  training shallow neural networks
Towards moderate overparameterization: global convergence guarantees for training shallow neural networksIEEE Journal on Selected Areas in Information Theory (JSAIT), 2019
Samet Oymak
Mahdi Soltanolkotabi
344
341
0
12 Feb 2019
Mean Field Limit of the Learning Dynamics of Multilayer Neural Networks
Mean Field Limit of the Learning Dynamics of Multilayer Neural Networks
Phan-Minh Nguyen
AI4CE
263
74
0
07 Feb 2019
Depth creates no more spurious local minima
Depth creates no more spurious local minima
Li Zhang
281
20
0
28 Jan 2019
On Connected Sublevel Sets in Deep Learning
On Connected Sublevel Sets in Deep Learning
Quynh N. Nguyen
429
106
0
22 Jan 2019
On the Benefit of Width for Neural Networks: Disappearance of Bad Basins
On the Benefit of Width for Neural Networks: Disappearance of Bad Basins
Dawei Li
Tian Ding
Tian Ding
706
42
0
28 Dec 2018
Overparameterized Nonlinear Learning: Gradient Descent Takes the
  Shortest Path?
Overparameterized Nonlinear Learning: Gradient Descent Takes the Shortest Path?
Samet Oymak
Mahdi Soltanolkotabi
ODL
360
194
0
25 Dec 2018
A jamming transition from under- to over-parametrization affects loss
  landscape and generalization
A jamming transition from under- to over-parametrization affects loss landscape and generalization
S. Spigler
Mario Geiger
Stéphane dÁscoli
Levent Sagun
Giulio Biroli
Matthieu Wyart
465
161
0
22 Oct 2018
On the loss landscape of a class of deep neural networks with no bad
  local valleys
On the loss landscape of a class of deep neural networks with no bad local valleysInternational Conference on Learning Representations (ICLR), 2018
Quynh N. Nguyen
Mahesh Chandra Mukkamala
Matthias Hein
518
91
0
27 Sep 2018
On the Global Convergence of Gradient Descent for Over-parameterized
  Models using Optimal Transport
On the Global Convergence of Gradient Descent for Over-parameterized Models using Optimal Transport
Lénaïc Chizat
Francis R. Bach
OT
614
818
0
24 May 2018
12
Next
Page 1 of 2