Learning theory of randomized Kaczmarz algorithm

Junhong Lin, Ding-Xuan Zhou*

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

30 Citations (Scopus)

Abstract

A relaxed randomized Kaczmarz algorithm is investigated in a least squares regression setting by a learning theory approach. When the sampling values are accurate and the regression function (conditional means) is linear, such an algorithm has been well studied in the community of non-uniform sampling. In this paper, we are mainly interested in the different case of either noisy random measurements or a nonlinear regression function. In this case, we show that relaxation is needed. A necessary and sufficient condition on the sequence of relaxation parameters or step sizes for the convergence of the algorithm in expectation is presented. Moreover, polynomial rates of convergence, both in expectation and in probability, are provided explicitly. As a result, the almost sure convergence of the algorithm is proved by applying the Borel-Cantelli Lemma.
Original languageEnglish
Pages (from-to)3341-3365
JournalJournal of Machine Learning Research (Print)
Volume16
Online publishedDec 2015
Publication statusPublished - 2015

Research Keywords

  • Almost sure convergence
  • Learning theory
  • Online learning
  • Relaxed randomized Kaczmarz algorithm
  • Space of homogeneous linear functions

Fingerprint

Dive into the research topics of 'Learning theory of randomized Kaczmarz algorithm'. Together they form a unique fingerprint.

Cite this