ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.07371
28
0

Minimum width for universal approximation using squashable activation functions

10 April 2025
Jonghyun Shin
Namjun Kim
Geonho Hwang
Sejun Park
ArXivPDFHTML
Abstract

The exact minimum width that allows for universal approximation of unbounded-depth networks is known only for ReLU and its variants. In this work, we study the minimum width of networks using general activation functions. Specifically, we focus on squashable functions that can approximate the identity function and binary step function by alternatively composing with affine transformations. We show that for networks using a squashable activation function to universally approximate LpL^pLp functions from [0,1]dx[0,1]^{d_x}[0,1]dx​ to Rdy\mathbb R^{d_y}Rdy​, the minimum width is max⁡{dx,dy,2}\max\{d_x,d_y,2\}max{dx​,dy​,2} unless dx=dy=1d_x=d_y=1dx​=dy​=1; the same bound holds for dx=dy=1d_x=d_y=1dx​=dy​=1 if the activation function is monotone. We then provide sufficient conditions for squashability and show that all non-affine analytic functions and a class of piecewise functions are squashable, i.e., our minimum width result holds for those general classes of activation functions.

View on arXiv
@article{shin2025_2504.07371,
  title={ Minimum width for universal approximation using squashable activation functions },
  author={ Jonghyun Shin and Namjun Kim and Geonho Hwang and Sejun Park },
  journal={arXiv preprint arXiv:2504.07371},
  year={ 2025 }
}
Comments on this paper