Factored Bandits

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

We introduce the factored bandits model, which is a framework for learning with limited (bandit) feedback, where actions can be decomposed into a Cartesian product of atomic actions. Factored bandits incorporate rank-1 bandits as a special case, but significantly relax the assumptions on the form of the reward function. We provide an anytime algorithm for stochastic factored bandits and up to constants matching upper and lower regret bounds for the problem. Furthermore, we show that with a slight modification the proposed algorithm can be applied to utility based dueling bandits. We obtain an improvement in the additive terms of the regret bound compared to state of the art algorithms (the additive terms are dominating up to time horizons which are exponential in the number of arms).
OriginalsprogEngelsk
TitelProceedings of 32nd Conference on Neural Information Processing Systems (NeurIPS 2018), Montréal, Canada.
Antal sider10
ForlagNIPS Proceedings
Publikationsdato2018
StatusUdgivet - 2018
Begivenhed32nd Annual Conference on Neural Information Processing Systems - Montreal, Montreal, Canada
Varighed: 2 dec. 20188 dec. 2018
Konferencens nummer: 32
https://nips.cc/Conferences/2018

Konference

Konference32nd Annual Conference on Neural Information Processing Systems
Nummer32
LokationMontreal
LandCanada
ByMontreal
Periode02/12/201808/12/2018
Internetadresse
NavnAdvances in Neural Information Processing Systems
Vol/bind31
ISSN1049-5258

ID: 225479776