Nonparametric Regression Using Over-parameterized Shallow ReLU Neural Networks

Yunfei Yang*, Ding-Xuan Zhou

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

5 Downloads (CityUHK Scholars)

Abstract

It is shown that over -parameterized neural networks can achieve minimax optimal rates of convergence (up to logarithmic factors) for learning functions from certain smooth function classes, if the weights are suitably constrained or regularized. Specifically, we consider the nonparametric regression of estimating an unknown d-variate function by using shallow ReLU neural networks. It is assumed that the regression function is from the Hölder space with smoothness α < (d+3)/2 or a variation space corresponding to shallow neural networks, which can be viewed as an infinitely wide neural network. In this setting, we prove that least squares estimators based on shallow neural networks with certain norm constraints on the weights are minimax optimal, if the network width is sufficiently large. As a byproduct, we derive a new size -independent bound for the local Rademacher complexity of shallow ReLU neural networks, which may be of independent interest. © 2024 Yunfei Yang and Ding-Xuan Zhou.
Original languageEnglish
Article number165
JournalJournal of Machine Learning Research
Volume25
Online published24 May 2024
Publication statusPublished - 2024

Funding

The work described in this paper was partially supported by InnoHK initiative, The Government of the HKSAR, Laboratory for AI-Powered Financial Technologies, the Research Grants Council of Hong Kong [Projects No. CityU 11306220 and 11308020] and National Natural Science Foundation of China [Project No. 12371103] when the second author worked at City University of Hong Kong. We thank the referees for their helpful comments and suggestions on the paper.

Research Keywords

  • neural networks
  • nonparametric regression
  • over-parameterization
  • regularization
  • rate of convergence

Publisher's Copyright Statement

  • This full text is made available under CC-BY 4.0. https://creativecommons.org/licenses/by/4.0/

Fingerprint

Dive into the research topics of 'Nonparametric Regression Using Over-parameterized Shallow ReLU Neural Networks'. Together they form a unique fingerprint.

Cite this