Double Doubly Robust Thompson Sampling for Generalized Linear Contextual Bandits
DOI:
https://doi.org/10.1609/aaai.v37i7.26001Keywords:
ML: Online Learning & Bandits, RU: Sequential Decision MakingAbstract
We propose a novel algorithm for generalized linear contextual bandits (GLBs) with a regret bound sublinear to the time horizon, the minimum eigenvalue of the covariance of contexts and a lower bound of the variance of rewards. In several identified cases, our result is the first regret bound for generalized linear bandits (GLBs) achieving the regret bound sublinear to the dimension of contexts without discarding the observed rewards. Previous approaches achieve the regret bound sublinear to the dimension of contexts by discarding the observed rewards, whereas our algorithm achieves the bound incorporating contexts from all arms in our double doubly robust (DDR) estimator. The DDR estimator is a subclass of doubly robust estimator but with a tighter error bound. We also provide a logarithmic cumulative regret bound under a probabilistic margin condition. This is the first regret bound under the margin condition for linear models or GLMs when contexts are different for all arms but coefficients are common. We conduct empirical studies using synthetic data and real examples, demonstrating the effectiveness of our algorithm.Downloads
Published
2023-06-26
How to Cite
Kim, W., Lee, K., & Paik, M. C. (2023). Double Doubly Robust Thompson Sampling for Generalized Linear Contextual Bandits. Proceedings of the AAAI Conference on Artificial Intelligence, 37(7), 8300-8307. https://doi.org/10.1609/aaai.v37i7.26001
Issue
Section
AAAI Technical Track on Machine Learning II