A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2019; you can also visit the original URL.
The file type is application/pdf
.
Greedy Confidence Pursuit: A Pragmatic Approach to Multi-bandit Optimization
[chapter]
2013
Lecture Notes in Computer Science
We address the practical problem of maximizing the number of high-confidence results produced among multiple experiments sharing an exhaustible pool of resources. We formalize this problem in the framework of bandit optimization as follows: given a set of multiple multi-armed bandits and a budget on the total number of trials allocated among them, select the top-m arms (with high confidence) for as many of the bandits as possible. To solve this problem, which we call greedy confidence pursuit,
doi:10.1007/978-3-642-40988-2_16
fatcat:3gp4ikabnvgvrjdaetq4x3th64