ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2309.10402
8
10

Minimum width for universal approximation using ReLU networks on compact domain

19 September 2023
Namjun Kim
Chanho Min
Sejun Park
    VLM
ArXivPDFHTML
Abstract

It has been shown that deep neural networks of a large enough width are universal approximators but they are not if the width is too small. There were several attempts to characterize the minimum width wmin⁡w_{\min}wmin​ enabling the universal approximation property; however, only a few of them found the exact values. In this work, we show that the minimum width for LpL^pLp approximation of LpL^pLp functions from [0,1]dx[0,1]^{d_x}[0,1]dx​ to Rdy\mathbb R^{d_y}Rdy​ is exactly max⁡{dx,dy,2}\max\{d_x,d_y,2\}max{dx​,dy​,2} if an activation function is ReLU-Like (e.g., ReLU, GELU, Softplus). Compared to the known result for ReLU networks, wmin⁡=max⁡{dx+1,dy}w_{\min}=\max\{d_x+1,d_y\}wmin​=max{dx​+1,dy​} when the domain is Rdx\smash{\mathbb R^{d_x}}Rdx​, our result first shows that approximation on a compact domain requires smaller width than on Rdx\smash{\mathbb R^{d_x}}Rdx​. We next prove a lower bound on wmin⁡w_{\min}wmin​ for uniform approximation using general activation functions including ReLU: wmin⁡≥dy+1w_{\min}\ge d_y+1wmin​≥dy​+1 if dx<dy≤2dxd_x<d_y\le2d_xdx​<dy​≤2dx​. Together with our first result, this shows a dichotomy between LpL^pLp and uniform approximations for general activation functions and input/output dimensions.

View on arXiv
Comments on this paper