Learning theory viewpoint of approximation by positive linear operators

Research output: Journal Publications and Reviews (RGC: 21, 22, 62)21_Publication in refereed journal

View graph of relations

Author(s)

  • Shaogao LV
  • Lei Shi

Detail(s)

Original languageEnglish
Pages (from-to)3177-3186
Journal / PublicationComputers and Mathematics with Applications
Volume60
Issue number12
Online published28 Oct 2010
Publication statusPublished - Dec 2010

Abstract

We follow a learning theory viewpoint to study a family of learning schemes for regression related to positive linear operators in approximation theory. Such a learning scheme is generated from a random sample by a kernel function parameterized by a scaling parameter. The essential difference between this algorithm and the classical approximation schemes is the randomness of the sampling points, which breaks the condition of good distribution of sampling points often required in approximation theory. We investigate the efficiency of the learning algorithm in a regression setting and present learning rates stated in terms of the smoothness of the regression function, sizes of variances, and distances of kernel centers from regular grids. The error analysis is conducted by estimating the sample error and the approximation error. Two examples with kernel functions related to continuous Bernstein bases and Jackson kernels are studied in detail and concrete learning rates are obtained.

Research Area(s)

  • Positive linear operator, Approximation theory, Regression function, Learning theory, Bernstein and Jackson kernels