Stochastic Continuous Submodular Maximization: Boosting via Non-oblivious Function

Qixin Zhang, Zengde Deng, Zaiyi Chen, Haoyuan Hu, Yu Yang*

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

21 Citations (Scopus)

Abstract

In this paper, we revisit Stochastic Continuous Submodular Maximization in both offline and online settings, which can benefit wide applications in machine learning and operations research areas. We present a boosting framework covering gradient ascent and online gradient ascent. The fundamental ingredient of our methods is a novel non-oblivious function F derived from a factor-revealing optimization problem, whose any stationary point provides a (1 - e-γ)-approximation to the global maximum of the γ-weakly DR-submodular objective function f ε C1,1(Χ). Under the offline scenario, we propose a boosting gradient ascent method achieving (1 - e-γε2)$-approximation after O(1/ε2) iterations, which improves the (γ2/1+γ2) approximation ratio of the classical gradient ascent algorithm. In the online setting, for the first time we consider the adversarial delays for stochastic gradient feedback, under which we propose a boosting online gradient algorithm with the same non-oblivious function F. Meanwhile, we verify that this boosting online algorithm achieves a regret of O(√D) against a (1 - e-γ)-approximation to the best feasible solution in hindsight, where D is the sum of delays of gradient feedback. To the best of our knowledge, this is the first result to obtain O(√T) regret against a (1 - e-γ)-approximation with O(1) gradient inquiry at each time step, when no delay exists, i.e., D = T. Finally, numerical experiments demonstrate the effectiveness of our boosting methods.

© 2022 by the author(s)
Original languageEnglish
Title of host publicationProceedings of the 39th International Conference on Machine Learning
EditorsKamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, Sivan Sabato
PublisherPMLR
Pages26116-26134
Number of pages19
Volume162
Publication statusPublished - Jul 2022
Event39th International Conference on Machine Learning (ICML 2022) - Hybrid, Baltimore, United States
Duration: 17 Jul 202223 Jul 2022
https://icml.cc/virtual/2022/index.html
https://icml.cc/Conferences/2022
https://proceedings.mlr.press/v162/

Publication series

NameProceedings of Machine Learning Research
PublisherPMLR
Volume162
ISSN (Print)2640-3498

Conference

Conference39th International Conference on Machine Learning (ICML 2022)
PlaceUnited States
CityBaltimore
Period17/07/2223/07/22
Internet address

Fingerprint

Dive into the research topics of 'Stochastic Continuous Submodular Maximization: Boosting via Non-oblivious Function'. Together they form a unique fingerprint.

Cite this