ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2110.02732
  4. Cited By
On Margin Maximization in Linear and ReLU Networks

On Margin Maximization in Linear and ReLU Networks

6 October 2021
Gal Vardi
Ohad Shamir
Nathan Srebro
ArXivPDFHTML

Papers citing "On Margin Maximization in Linear and ReLU Networks"

6 / 6 papers shown
Title
The late-stage training dynamics of (stochastic) subgradient descent on homogeneous neural networks
Sholom Schechtman
Nicolas Schreuder
50
0
0
08 Feb 2025
The Implicit Bias of Batch Normalization in Linear Models and Two-layer
  Linear Convolutional Neural Networks
The Implicit Bias of Batch Normalization in Linear Models and Two-layer Linear Convolutional Neural Networks
Yuan Cao
Difan Zou
Yuan-Fang Li
Quanquan Gu
MLT
22
5
0
20 Jun 2023
On the Implicit Bias in Deep-Learning Algorithms
On the Implicit Bias in Deep-Learning Algorithms
Gal Vardi
FedML
AI4CE
22
72
0
26 Aug 2022
Reconstructing Training Data from Trained Neural Networks
Reconstructing Training Data from Trained Neural Networks
Niv Haim
Gal Vardi
Gilad Yehudai
Ohad Shamir
Michal Irani
16
130
0
15 Jun 2022
On the Effective Number of Linear Regions in Shallow Univariate ReLU
  Networks: Convergence Guarantees and Implicit Bias
On the Effective Number of Linear Regions in Shallow Univariate ReLU Networks: Convergence Guarantees and Implicit Bias
Itay Safran
Gal Vardi
Jason D. Lee
MLT
41
23
0
18 May 2022
Towards Understanding Learning in Neural Networks with Linear Teachers
Towards Understanding Learning in Neural Networks with Linear Teachers
Roei Sarussi
Alon Brutzkus
Amir Globerson
FedML
MLT
44
20
0
07 Jan 2021
1