Distributed Kernel-Based Gradient Descent Algorithms

Shao-Bo Lin*, Ding-Xuan Zhou

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

71 Citations (Scopus)

Abstract

We study the generalization ability of distributed learning equipped with a divide-and-conquer approach and gradient descent algorithm in a reproducing kernel Hilbert space (RKHS). Using special spectral features of the gradient descent algorithms and a novel integral operator approach, we provide optimal learning rates of distributed gradient descent algorithms in probability and partly conquer the saturation phenomenon in the literature in the sense that the maximum number of local machines to guarantee the optimal learning rates does not vary if the regularity of the regression function goes beyond a certain quantity. We also find that additional unlabeled data can help relax the restriction on the number of local machines in distributed learning.
Original languageEnglish
Pages (from-to)249-276
JournalConstructive Approximation
Volume47
Issue number2
Online published15 May 2017
DOIs
Publication statusPublished - Apr 2018

Research Keywords

  • Distributed learning
  • Gradient descent algorithm
  • Integral operator
  • Learning theory

Fingerprint

Dive into the research topics of 'Distributed Kernel-Based Gradient Descent Algorithms'. Together they form a unique fingerprint.

Cite this