A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2022; you can also visit the original URL.
The file type is application/pdf
.
Disposable Linear Bandits for Online Recommendations
2021
AAAI Conference on Artificial Intelligence
We study the classic stochastic linear bandit problem under the restriction that each arm may be selected for limited number of times. This simple constraint, which we call disposability, captures a common restriction that occurs in recommendation problems from a diverse array of applications ranging from personalized styling services to dating platforms. We show that the regret for this problem is characterized by a previously-unstudied function of the reward distribution among optimal arms.
dblp:conf/aaai/KorkutL21
fatcat:f2mefsizdnfijh7es6dvh23pbq