ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2103.09177
  4. Cited By
Deep learning: a statistical viewpoint

Deep learning: a statistical viewpoint

16 March 2021
Peter L. Bartlett
Andrea Montanari
Alexander Rakhlin
ArXiv (abs)PDFHTML

Papers citing "Deep learning: a statistical viewpoint"

43 / 93 papers shown
Title
Deep Generative Modeling on Limited Data with Regularization by
  Nontransferable Pre-trained Models
Deep Generative Modeling on Limited Data with Regularization by Nontransferable Pre-trained Models
Yong Zhong
Hongtao Liu
Xiaodong Liu
Fan Bao
Weiran Shen
Chongxuan Li
AI4CE
90
4
0
30 Aug 2022
Heterogeneous Treatment Effect with Trained Kernels of the
  Nadaraya-Watson Regression
Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya-Watson Regression
A. Konstantinov
Stanislav R. Kirpichenko
Lev V. Utkin
CML
54
4
0
19 Jul 2022
Research Trends and Applications of Data Augmentation Algorithms
Research Trends and Applications of Data Augmentation Algorithms
João Fonseca
F. Bação
45
4
0
18 Jul 2022
Tensor-on-Tensor Regression: Riemannian Optimization,
  Over-parameterization, Statistical-computational Gap, and Their Interplay
Tensor-on-Tensor Regression: Riemannian Optimization, Over-parameterization, Statistical-computational Gap, and Their Interplay
Yuetian Luo
Anru R. Zhang
106
20
0
17 Jun 2022
Benefits of Additive Noise in Composing Classes with Bounded Capacity
Benefits of Additive Noise in Composing Classes with Bounded Capacity
A. F. Pour
H. Ashtiani
66
3
0
14 Jun 2022
Invariant Structure Learning for Better Generalization and Causal
  Explainability
Invariant Structure Learning for Better Generalization and Causal Explainability
Yunhao Ge
Sercan O. Arik
Jinsung Yoon
Ao Xu
Laurent Itti
Tomas Pfister
OODCML
49
2
0
13 Jun 2022
The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at
  Initialization
The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at Initialization
Mufan Li
Mihai Nica
Daniel M. Roy
104
39
0
06 Jun 2022
Generalization for multiclass classification with overparameterized
  linear models
Generalization for multiclass classification with overparameterized linear models
Vignesh Subramanian
Rahul Arya
A. Sahai
AI4CE
73
9
0
03 Jun 2022
VC Theoretical Explanation of Double Descent
VC Theoretical Explanation of Double Descent
Eng Hock Lee
V. Cherkassky
57
3
0
31 May 2022
Blind Estimation of a Doubly Selective OFDM Channel: A Deep Learning
  Algorithm and Theory
Blind Estimation of a Doubly Selective OFDM Channel: A Deep Learning Algorithm and Theory
T. Getu
N. Golmie
D. Griffith
52
2
0
30 May 2022
Surprises in adversarially-trained linear regression
Surprises in adversarially-trained linear regression
Antônio H. Ribeiro
Dave Zachariah
Thomas B. Schon
AAML
184
2
0
25 May 2022
High-dimensional Asymptotics of Feature Learning: How One Gradient Step
  Improves the Representation
High-dimensional Asymptotics of Feature Learning: How One Gradient Step Improves the Representation
Jimmy Ba
Murat A. Erdogdu
Taiji Suzuki
Zhichao Wang
Denny Wu
Greg Yang
MLT
99
129
0
03 May 2022
The Directional Bias Helps Stochastic Gradient Descent to Generalize in
  Kernel Regression Models
The Directional Bias Helps Stochastic Gradient Descent to Generalize in Kernel Regression Models
Yiling Luo
X. Huo
Y. Mei
44
0
0
29 Apr 2022
Flexible Multiple-Objective Reinforcement Learning for Chip Placement
Flexible Multiple-Objective Reinforcement Learning for Chip Placement
Fu-Chieh Chang
Yu-Wei Tseng
Ya-Wen Yu
Ssu-Rui Lee
Alexandru Cioba
...
Chien-Yi Yang
Ren-Chu Wang
Yao-Wen Chang
Tai-Chen Chen
Tung-Chieh Chen
OffRL
60
5
0
13 Apr 2022
Convergence of gradient descent for deep neural networks
Convergence of gradient descent for deep neural networks
S. Chatterjee
ODL
63
22
0
30 Mar 2022
Coarsening the Granularity: Towards Structurally Sparse Lottery Tickets
Coarsening the Granularity: Towards Structurally Sparse Lottery Tickets
Tianlong Chen
Xuxi Chen
Xiaolong Ma
Yanzhi Wang
Zhangyang Wang
82
34
0
09 Feb 2022
Is interpolation benign for random forest regression?
Is interpolation benign for random forest regression?
Ludovic Arnould
Claire Boyer
Erwan Scornet
66
6
0
08 Feb 2022
Datamodels: Predicting Predictions from Training Data
Datamodels: Predicting Predictions from Training Data
Andrew Ilyas
Sung Min Park
Logan Engstrom
Guillaume Leclerc
Aleksander Madry
TDI
135
143
0
01 Feb 2022
Nearest Class-Center Simplification through Intermediate Layers
Nearest Class-Center Simplification through Intermediate Layers
Ido Ben-Shaul
S. Dekel
95
27
0
21 Jan 2022
On generalization bounds for deep networks based on loss surface
  implicit regularization
On generalization bounds for deep networks based on loss surface implicit regularization
Masaaki Imaizumi
Johannes Schmidt-Hieber
ODL
68
3
0
12 Jan 2022
Causal Discovery from Sparse Time-Series Data Using Echo State Network
Causal Discovery from Sparse Time-Series Data Using Echo State Network
Haonan Chen
B. Chang
Mohamed A. Naiel1
Georges Younes
Steven Wardell
Stan Kleinikkink
J. S. Z. U. O. Waterloo
CMLBDLAI4TS
50
1
0
09 Jan 2022
Measuring Complexity of Learning Schemes Using Hessian-Schatten Total
  Variation
Measuring Complexity of Learning Schemes Using Hessian-Schatten Total Variation
Shayan Aziznejad
Joaquim Campos
M. Unser
93
10
0
12 Dec 2021
On the Convergence of Shallow Neural Network Training with Randomly
  Masked Neurons
On the Convergence of Shallow Neural Network Training with Randomly Masked Neurons
Fangshuo Liao
Anastasios Kyrillidis
118
16
0
05 Dec 2021
Harmless interpolation in regression and classification with structured
  features
Harmless interpolation in regression and classification with structured features
Andrew D. McRae
Santhosh Karnik
Mark A. Davenport
Vidya Muthukumar
186
11
0
09 Nov 2021
Binary perceptron: efficient algorithms can find solutions in a rare
  well-connected cluster
Binary perceptron: efficient algorithms can find solutions in a rare well-connected cluster
Emmanuel Abbe
Shuangping Li
Allan Sly
MQ
77
33
0
04 Nov 2021
Mean-field Analysis of Piecewise Linear Solutions for Wide ReLU Networks
Mean-field Analysis of Piecewise Linear Solutions for Wide ReLU Networks
Aleksandr Shevchenko
Vyacheslav Kungurtsev
Marco Mondelli
MLT
93
13
0
03 Nov 2021
Does the Data Induce Capacity Control in Deep Learning?
Does the Data Induce Capacity Control in Deep Learning?
Rubing Yang
Jialin Mao
Pratik Chaudhari
124
16
0
27 Oct 2021
Data splitting improves statistical performance in overparametrized
  regimes
Data splitting improves statistical performance in overparametrized regimes
Nicole Mücke
Enrico Reiss
Jonas Rungenhagen
Markus Klein
53
8
0
21 Oct 2021
The Neural Testbed: Evaluating Joint Predictions
The Neural Testbed: Evaluating Joint Predictions
Ian Osband
Zheng Wen
S. Asghari
Vikranth Dwaracherla
Botao Hao
M. Ibrahimi
Dieterich Lawson
Xiuyuan Lu
Brendan O'Donoghue
Benjamin Van Roy
UQCV
94
22
0
09 Oct 2021
VC dimension of partially quantized neural networks in the
  overparametrized regime
VC dimension of partially quantized neural networks in the overparametrized regime
Yutong Wang
Clayton D. Scott
81
1
0
06 Oct 2021
Deformed semicircle law and concentration of nonlinear random matrices
  for ultra-wide neural networks
Deformed semicircle law and concentration of nonlinear random matrices for ultra-wide neural networks
Zhichao Wang
Yizhe Zhu
104
20
0
20 Sep 2021
A Farewell to the Bias-Variance Tradeoff? An Overview of the Theory of
  Overparameterized Machine Learning
A Farewell to the Bias-Variance Tradeoff? An Overview of the Theory of Overparameterized Machine Learning
Yehuda Dar
Vidya Muthukumar
Richard G. Baraniuk
117
72
0
06 Sep 2021
Identifiable Energy-based Representations: An Application to Estimating
  Heterogeneous Causal Effects
Identifiable Energy-based Representations: An Application to Estimating Heterogeneous Causal Effects
Yao Zhang
Jeroen Berrevoets
M. Schaar
CML
84
6
0
06 Aug 2021
Stability & Generalisation of Gradient Descent for Shallow Neural
  Networks without the Neural Tangent Kernel
Stability & Generalisation of Gradient Descent for Shallow Neural Networks without the Neural Tangent Kernel
Dominic Richards
Ilja Kuzborskij
79
29
0
27 Jul 2021
A brief note on understanding neural networks as Gaussian processes
A brief note on understanding neural networks as Gaussian processes
Mengwu Guo
BDLGP
81
2
0
25 Jul 2021
Nonparametric Regression with Shallow Overparameterized Neural Networks Trained by GD with Early Stopping
Ilja Kuzborskij
Csaba Szepesvári
98
7
0
12 Jul 2021
A Generalized Lottery Ticket Hypothesis
A Generalized Lottery Ticket Hypothesis
Ibrahim Alabdulmohsin
L. Markeeva
Daniel Keysers
Ilya O. Tolstikhin
56
6
0
03 Jul 2021
Fit without fear: remarkable mathematical phenomena of deep learning
  through the prism of interpolation
Fit without fear: remarkable mathematical phenomena of deep learning through the prism of interpolation
M. Belkin
70
186
0
29 May 2021
Universal Prediction Band via Semi-Definite Programming
Universal Prediction Band via Semi-Definite Programming
Tengyuan Liang
67
3
0
31 Mar 2021
The Low-Rank Simplicity Bias in Deep Networks
The Low-Rank Simplicity Bias in Deep Networks
Minyoung Huh
H. Mobahi
Richard Y. Zhang
Brian Cheung
Pulkit Agrawal
Phillip Isola
117
116
0
18 Mar 2021
Benign overfitting in ridge regression
Benign overfitting in ridge regression
Alexander Tsigler
Peter L. Bartlett
91
169
0
29 Sep 2020
The Interpolation Phase Transition in Neural Networks: Memorization and
  Generalization under Lazy Training
The Interpolation Phase Transition in Neural Networks: Memorization and Generalization under Lazy Training
Andrea Montanari
Yiqiao Zhong
190
97
0
25 Jul 2020
Surprises in High-Dimensional Ridgeless Least Squares Interpolation
Surprises in High-Dimensional Ridgeless Least Squares Interpolation
Trevor Hastie
Andrea Montanari
Saharon Rosset
Robert Tibshirani
291
747
0
19 Mar 2019
Previous
12