Efficient Projection-Free Online Methods with Stochastic Recursive Gradient

Authors

  • Jiahao Xie Zhejiang University
  • Zebang Shen University of Pennsylvania
  • Chao Zhang Zhejiang University
  • Boyu Wang University of Western Ontario
  • Hui Qian Zhejiang University

DOI:

https://doi.org/10.1609/aaai.v34i04.6116

Abstract

This paper focuses on projection-free methods for solving smooth Online Convex Optimization (OCO) problems. Existing projection-free methods either achieve suboptimal regret bounds or have high per-round computational costs. To fill this gap, two efficient projection-free online methods called ORGFW and MORGFW are proposed for solving stochastic and adversarial OCO problems, respectively. By employing a recursive gradient estimator, our methods achieve optimal regret bounds (up to a logarithmic factor) while possessing low per-round computational costs. Experimental results demonstrate the efficiency of the proposed methods compared to state-of-the-arts.

Downloads

Published

2020-04-03

How to Cite

Xie, J., Shen, Z., Zhang, C., Wang, B., & Qian, H. (2020). Efficient Projection-Free Online Methods with Stochastic Recursive Gradient. Proceedings of the AAAI Conference on Artificial Intelligence, 34(04), 6446-6453. https://doi.org/10.1609/aaai.v34i04.6116

Issue

Section

AAAI Technical Track: Machine Learning