Approximation of Nonlinear Functionals Using Deep ReLU Networks
Research output: Journal Publications and Reviews (RGC: 21, 22, 62) › 21_Publication in refereed journal › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Article number | 50 |
Journal / Publication | Journal of Fourier Analysis and Applications |
Volume | 29 |
Issue number | 4 |
Online published | 28 Jul 2023 |
Publication status | Published - Aug 2023 |
Link(s)
Abstract
In recent years, functional neural networks have been proposed and studied in order to approximate nonlinear continuous functionals defined on Lp([- 1 , 1] s) for integers s≥ 1 and 1 ≤ p< ∞ . However, their theoretical properties are largely unknown beyond universality of approximation or the existing analysis does not apply to the rectified linear unit (ReLU) activation function. To fill in this void, we investigate here the approximation power of functional deep neural networks associated with the ReLU activation function by constructing a continuous piecewise linear interpolation under a simple triangulation. In addition, we establish rates of approximation of the proposed functional deep ReLU networks under mild regularity conditions. Finally, our study may also shed some light on the understanding of functional data learning algorithms. © 2023, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
Research Area(s)
- Approximation theory, Deep learning theory, Functional neural networks, Modulus of continuity, ReLU
Citation Format(s)
Approximation of Nonlinear Functionals Using Deep ReLU Networks. / Song, Linhao; Fan, Jun; Chen, Di-Rong et al.
In: Journal of Fourier Analysis and Applications, Vol. 29, No. 4, 50, 08.2023.
In: Journal of Fourier Analysis and Applications, Vol. 29, No. 4, 50, 08.2023.
Research output: Journal Publications and Reviews (RGC: 21, 22, 62) › 21_Publication in refereed journal › peer-review