Abstract
Efficiently trading off exploration and exploitation is one of the key challenges in online Reinforcement Learning (RL). Most works achieve this by carefully estimating the model uncertainty and following the so-called optimistic model. Inspired by practical ensemble methods, in this work we propose a simple and novel batch ensemble scheme that provably achieves near-optimal regret for stochastic Multi-Armed Bandits (MAB). Crucially, our algorithm has just a single parameter, namely the number of batches, and its value does not depend on distributional properties such as the scale and variance of the losses. We complement our theoretical results by demonstrating the effectiveness of our algorithm on synthetic benchmarks.
Original language | English |
---|---|
Pages (from-to) | 15678-15685 |
Number of pages | 8 |
Journal | Proceedings of the AAAI Conference on Artificial Intelligence |
Volume | 39 |
Issue number | 15 |
DOIs | |
State | Published - 11 Apr 2025 |
Event | 39th Annual AAAI Conference on Artificial Intelligence, AAAI 2025 - Philadelphia, United States Duration: 25 Feb 2025 → 4 Mar 2025 |
Funding
Funders | Funder number |
---|---|
Yandex Initiative for Machine Learning | |
Blavatnik Family Foundation | |
Tel Aviv University | |
European Research Council | |
Horizon 2020 Framework Programme | 882396, 101078075 |
Israel Science Foundation | 2549/19, 993/17 |