Deep neural networks for rotation-invariance approximation and learning
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Pages (from-to) | 737-772 |
Journal / Publication | Analysis and Applications |
Volume | 17 |
Issue number | 5 |
Online published | 15 Aug 2019 |
Publication status | Published - Sept 2019 |
Link(s)
Abstract
Based on the tree architecture, the objective of this paper is to design deep neural networks with two or more hidden layers (called deep nets) for realization of radial functions so as to enable rotational invariance for near-optimal function approximation in an arbitrarily high-dimensional Euclidian space. It is shown that deep nets have much better performance than shallow nets (with only one hidden layer) in terms of approximation accuracy and learning capabilities. In particular, for learning radial functions, it is shown that near-optimal rate can be achieved by deep nets but not by shallow nets. Our results illustrate the necessity of depth in neural network design for realization of rotation-invariance target functions.
Research Area(s)
- Deep nets, learning theory, radial-basis functions, rotation-invariance
Citation Format(s)
Deep neural networks for rotation-invariance approximation and learning. / Chui, Charles K.; Lin, Shao-Bo; Zhou, Ding-Xuan.
In: Analysis and Applications, Vol. 17, No. 5, 09.2019, p. 737-772.
In: Analysis and Applications, Vol. 17, No. 5, 09.2019, p. 737-772.
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review