Toward a data efficient neural actor-critic - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Communication Dans Un Congrès Année : 2016

Toward a data efficient neural actor-critic

Yann Boniface
  • Fonction : Auteur
  • PersonId : 778629
  • IdRef : 140410805
Alain Dutech

Résumé

A new off-policy, offline, model-free, actor-critic reinforcement learning algorithm dealing with continuous environments in both states and actions is presented. It addresses discrete time problems where the goal is to maximize the discounted sum of rewards using stationary policies. Our algorithm allows to trade-off between data-efficiency and scalability. The amount of a priori knowledge is kept low by: (1) using neural networks to learn both the critic and the actor, (2) not relying on initial trajectories provided by an expert, and (3) not depending on known goal states. Experimental results show better data-efficiency than 4 state-of-the-art algorithms on two benchmark environments.
Fichier principal
Vignette du fichier
ewrl13-2016-submission_7.pdf (783.25 Ko) Télécharger le fichier
Origine : Fichiers éditeurs autorisés sur une archive ouverte
Loading...

Dates et versions

hal-01413885 , version 1 (11-12-2016)

Identifiants

  • HAL Id : hal-01413885 , version 1

Citer

Matthieu Zimmer, Yann Boniface, Alain Dutech. Toward a data efficient neural actor-critic. EWRL 2016 - The 13th European Workshop on Reinforcement Learning, Dec 2016, Barcelona, Spain. ⟨hal-01413885⟩
354 Consultations
210 Téléchargements

Partager

Gmail Facebook X LinkedIn More