Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2010.14495
Cited By
Are wider nets better given the same number of parameters?
27 October 2020
A. Golubeva
Behnam Neyshabur
Guy Gur-Ari
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Are wider nets better given the same number of parameters?"
9 / 9 papers shown
Title
Learning to Manipulate under Limited Information
Wesley H. Holliday
Alexander Kristoffersen
Eric Pacuit
15
4
0
29 Jan 2024
REDS: Resource-Efficient Deep Subnetworks for Dynamic Resource Constraints
Francesco Corti
Balz Maag
Joachim Schauer
U. Pferschy
O. Saukh
26
2
0
22 Nov 2023
ReLU Neural Networks with Linear Layers are Biased Towards Single- and Multi-Index Models
Suzanna Parkinson
Greg Ongie
Rebecca Willett
60
6
0
24 May 2023
Sparse-IFT: Sparse Iso-FLOP Transformations for Maximizing Training Efficiency
Vithursan Thangarasa
Shreyas Saxena
Abhay Gupta
Sean Lie
25
3
0
21 Mar 2023
Deep Learning Meets Sparse Regularization: A Signal Processing Perspective
Rahul Parhi
Robert D. Nowak
21
25
0
23 Jan 2023
Understanding the effect of sparsity on neural networks robustness
Lukas Timpl
R. Entezari
Hanie Sedghi
Behnam Neyshabur
O. Saukh
31
11
0
22 Jun 2022
What Kinds of Functions do Deep Neural Networks Learn? Insights from Variational Spline Theory
Rahul Parhi
Robert D. Nowak
MLT
27
70
0
07 May 2021
Scaling Laws for Neural Language Models
Jared Kaplan
Sam McCandlish
T. Henighan
Tom B. Brown
B. Chess
R. Child
Scott Gray
Alec Radford
Jeff Wu
Dario Amodei
228
4,460
0
23 Jan 2020
Aggregated Residual Transformations for Deep Neural Networks
Saining Xie
Ross B. Girshick
Piotr Dollár
Z. Tu
Kaiming He
297
10,216
0
16 Nov 2016
1