Diversity-driven selection of exploration strategies in multi-armed bandits - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Communication Dans Un Congrès Année : 2015

Diversity-driven selection of exploration strategies in multi-armed bandits

Résumé

We consider a scenario where an agent has multiple available strategies to explore an unknown environment. For each new interaction with the environment, the agent must select which exploration strategy to use. We provide a new strategy-agnostic method that treat the situation as a Multi-Armed Bandits problem where the reward signal is the diversity of effects that each strategy produces. We test the method empirically on a simulated planar robotic arm, and establish that the method is both able discriminate between strategies of dissimilar quality, even when the differences are tenuous, and that the resulting performance is competitive with the best fixed mixture of strategies.
Fichier principal
Vignette du fichier
BenureauOudeyerIcdl2015.pdf (1.93 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01251060 , version 1 (05-01-2016)

Identifiants

Citer

Fabien Benureau, Pierre-Yves Oudeyer. Diversity-driven selection of exploration strategies in multi-armed bandits. IEEE International Conference on Development and Learning and Epigenetic Robotics, Aug 2015, Providence, United States. ⟨10.1109/DEVLRN.2015.7346130⟩. ⟨hal-01251060⟩

Collections

ENSTA INRIA INRIA2
137 Consultations
166 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More