On the Asymptotic Optimality of Confidence Interval Based Algorithms for Fixed Confidence MABs

Authors

  • Kushal Kejriwal Indian Institute of Technology Bombay
  • Nikhil Karamchandani Indian Institute of Technology Bombay
  • Jayakrishnan Nair Indian Institute of Technology Bombay

DOI:

https://doi.org/10.1609/aaai.v39i17.33959

Abstract

In this work, we address the challenge of identifying the optimal arm in a stochastic multi-armed bandit scenario with the minimum number of arm pulls, given a predefined error probability in a fixed confidence setting. Our focus is on examining the asymptotic behavior of sample complexity and the distribution of arm weights upon termination, as the error threshold is scaled to zero, under confidence-interval based algorithms. Specifically, we analyze the asymptotic sample complexity and termination weight fractions for the well-known LUCB algorithm, and introduce a new variant, the LUCB Greedy algorithm. We demonstrate that the upper bounds on the sample complexities for both algorithms are asymptotically within a constant factor of the established lower bounds.

Downloads

Published

2025-04-11

How to Cite

Kejriwal, K., Karamchandani, N., & Nair, J. (2025). On the Asymptotic Optimality of Confidence Interval Based Algorithms for Fixed Confidence MABs. Proceedings of the AAAI Conference on Artificial Intelligence, 39(17), 17814–17821. https://doi.org/10.1609/aaai.v39i17.33959

Issue

Section

AAAI Technical Track on Machine Learning III