Debiased Distributed Learning for Sparse Partial Linear Models in High Dimensions

Shaogao Lv, Heng Lian

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

13 Citations (Scopus)
37 Downloads (CityUHK Scholars)

Abstract

Although various distributed machine learning schemes have been proposed recently for purely linear models and fully nonparametric models, little attention has been paid to distributed optimization for semi-parametric models with multiple structures (e.g. sparsity, linearity and nonlinearity). To address these issues, the current paper proposes a new communication-efficient distributed learning algorithm for sparse partially linear models with an increasing number of features. The proposed method is based on the classical divide and conquer strategy for handling big data and the computation on each subsample consists of a debiased estimation of the doubly regularized least squares approach. With the proposed method, we theoretically prove that our global parametric estimator can achieve the optimal parametric rate in our semi-parametric model given an appropriate partition on the total data. Specifically, the choice of data partition relies on the underlying smoothness of the nonparametric component, and it is adaptive to the sparsity parameter. Finally, some simulated experiments are carried out to illustrate the empirical performances of our debiased technique under the distributed setting.
Original languageEnglish
Article number2
JournalJournal of Machine Learning Research
Volume23
Online publishedDec 2021
Publication statusPublished - 2022

Research Keywords

  • Big data
  • Distributed learning
  • High dimensions
  • Reproducing kernel Hilbert space (RKHS)
  • Semi-parametric models

Publisher's Copyright Statement

  • This full text is made available under CC-BY 4.0. https://creativecommons.org/licenses/by/4.0/

Fingerprint

Dive into the research topics of 'Debiased Distributed Learning for Sparse Partial Linear Models in High Dimensions'. Together they form a unique fingerprint.

Cite this