OAK

MinimumWidth for Universal Approximation using Squashable Activation Functions

Metadata Downloads
Author(s)
Shin, JonghyunKim, NamjunHwang, GeonhoPark, Sejun
Type
Conference Paper
Citation
42nd International Conference on Machine Learning-ICML-Annual, pp.55096 - 55121
Issued Date
2025-07-13
Abstract
The exact minimum width that allows for universal approximation of unbounded-depth networks is known only for RELU and its variants. In this work, we study the minimum width of networks using general activation functions. Specifically, we focus on squashable functions that can approximate the identity function and binary step function by alternatively composing with affine transformations. We show that for networks using a squashable activation function to universally approximate L-p functions from [0, 1](dx) to R-dy, the minimum width is max{d(x), d(y), 2} unless d(x) = d(y) = 1; the same bound holds for d(x) = d(y) = 1 if the activation function is monotone. We then provide sufficient conditions for squashability and show that all non-affine analytic functions and a class of piecewise functions are squashable, i.e., our minimum width result holds for those general classes of activation functions.
Publisher
ICML 2025
Conference Place
CN
URI
https://scholar.gist.ac.kr/handle/local/34016
공개 및 라이선스
  • 공개 구분공개
파일 목록
  • 관련 파일이 존재하지 않습니다.

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.