ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2303.11454
  4. Cited By
How (Implicit) Regularization of ReLU Neural Networks Characterizes the
  Learned Function -- Part II: the Multi-D Case of Two Layers with Random First
  Layer

How (Implicit) Regularization of ReLU Neural Networks Characterizes the Learned Function -- Part II: the Multi-D Case of Two Layers with Random First Layer

20 March 2023
Jakob Heiss
Josef Teichmann
Hanna Wutte
    AI4CE
ArXiv (abs)PDFHTML

Papers citing "How (Implicit) Regularization of ReLU Neural Networks Characterizes the Learned Function -- Part II: the Multi-D Case of Two Layers with Random First Layer"

1 / 1 papers shown
Title
Extending Path-Dependent NJ-ODEs to Noisy Observations and a Dependent
  Observation Framework
Extending Path-Dependent NJ-ODEs to Noisy Observations and a Dependent Observation Framework
William Andersson
Jakob Heiss
Florian Krach
Josef Teichmann
137
4
0
24 Jul 2023
1