Learning efficient exploration strategies from experience for Best Arm Identification
Determine whether efficient exploration strategies for Best Arm Identification in multi-armed bandit problems can be learned directly from experience, thereby avoiding the explicit design of instance-dependent Best Arm Identification algorithms.
References
Therefore, in this work we address the open question of whether it is possible to learn efficient exploration strategies directly from experience, avoiding the process of designing a BAI algorithm.
— Learning to Explore: An In-Context Learning Approach for Pure Exploration
(2506.01876 - Russo et al., 2 Jun 2025) in Section 2, Learning to Explore: In-Context Pure Exploration (paragraph on Best-Arm Identification)