Profitable Bandits - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Article Dans Une Revue Proceedings of Machine Learning Research Année : 2018

Profitable Bandits

Résumé

Originally motivated by default risk management applications, this paper investigates a novel problem, referred to as the profitable bandit problem here. At each step, an agent chooses a subset of the K ≥ 1 possible actions. For each action chosen, she then respectively pays and receives the sum of a random number of costs and rewards. Her objective is to maximize her cumulated profit. We adapt and study three well-known strategies in this purpose, that were proved to be most efficient in other settings: kl-UCB, Bayes-UCB and Thompson Sampling. For each of them, we prove a finite time regret bound which, together with a lower bound we obtain as well, establishes asymptotic optimality in some cases. Our goal is also to compare these three strategies from a theoretical and empirical perspective both at the same time. We give simple, self-contained proofs that emphasize their similarities, as well as their differences. While both Bayesian strategies are automatically adapted to the geometry of information, the numerical experiments carried out show a slight advantage for Thompson Sampling in practice.
Fichier principal
Vignette du fichier
main_acml.pdf (355.12 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-02023057 , version 1 (18-02-2019)

Identifiants

Citer

Mastane Achab, Stéphan Clémençon, Aurélien Garivier. Profitable Bandits. Proceedings of Machine Learning Research, 2018, 95, pp.694-709. ⟨hal-02023057⟩
80 Consultations
31 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More