Training Dialogue Systems With Human Advice - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Communication Dans Un Congrès Année : 2018

Training Dialogue Systems With Human Advice

Merwan Barlier
  • Fonction : Auteur
  • PersonId : 1040020
Romain Laroche
  • Fonction : Auteur
  • PersonId : 1012067

Résumé

One major drawback of Reinforcement Learning (RL) Spoken Dialogue Systems is that they inherit from the general exploration requirements of RL which makes them hard to deploy from an industry perspective. On the other hand, industrial systems rely on human expertise and hand written rules so as to avoid irrelevant behavior to happen and maintain acceptable experience from the user point of view. In this paper, we attempt to bridge the gap between those two worlds by providing an easy way to incorporate all kinds of human expertise in the training phase of a Reinforcement Learning Dialogue System. Our approach, based on the TAMER framework, enables safe and efficient policy learning by combining the traditional Reinforcement Learning reward signal with an additional reward, encoding expert advice. Experimental results show that our method leads to substantial improvements over more traditional Reinforcement Learning methods.
Fichier principal
Vignette du fichier
1-4-4-AAMAS-Training-Dialogue-Systems-with-Human-Advice.pdf (693.05 Ko) Télécharger le fichier
Origine : Fichiers éditeurs autorisés sur une archive ouverte
Loading...

Dates et versions

hal-01945831 , version 1 (11-12-2018)

Identifiants

  • HAL Id : hal-01945831 , version 1

Citer

Merwan Barlier, Romain Laroche, Olivier Pietquin. Training Dialogue Systems With Human Advice. AAMAS 2018 - the 17th International Conference on Autonomous Agents and Multiagent Systems, Jul 2018, Stockholm, Sweden. pp.9. ⟨hal-01945831⟩
66 Consultations
132 Téléchargements

Partager

Gmail Facebook X LinkedIn More