Deutsch
 
Hilfe Datenschutzhinweis Impressum
  DetailsucheBrowse

Datensatz

 
 
DownloadE-Mail
  Reinforcement Learning by Relative Entropy Policy Search

Peters, J., Mülling, K., & Altun, Y. (2010). Reinforcement Learning by Relative Entropy Policy Search.

Item is

Externe Referenzen

einblenden:

Urheber

einblenden:
ausblenden:
 Urheber:
Peters, J1, 2, Autor           
Mülling, K1, Autor           
Altun, Y1, Autor           
Affiliations:
1Department Empirical Inference, Max Planck Institute for Biological Cybernetics, Max Planck Society, ou_1497795              
2Dept. Empirical Inference, Max Planck Institute for Intelligent Systems, Max Planck Society, ou_1497647              

Inhalt

einblenden:
ausblenden:
Schlagwörter: -
 Zusammenfassung: Policy search is a successful approach to reinforcement learning. However, policy improvements often result in the loss of information. Hence, it has been marred by premature convergence and implausible solutions. As first suggested in the context of covariant policy gradients, many of these problems may be addressed by constraining the information loss. In this book chapter, we continue this path of reasoning and suggest the Relative Entropy Policy Search (REPS) method. The resulting method differs significantly from previous policy gradient approaches and yields an exact update step. It works well on typical reinforcement learning benchmark problems. We will also present a real-world applications where a robot employs REPS to learn how to return balls in a game of table tennis.

Details

einblenden:
ausblenden:
Sprache(n):
 Datum: 2010-07
 Publikationsstatus: Erschienen
 Seiten: -
 Ort, Verlag, Ausgabe: -
 Inhaltsverzeichnis: -
 Art der Begutachtung: -
 Identifikatoren: URI: http://maxent2010.inrialpes.fr/files/2010/06/booklet.pdf
BibTex Citekey: 6746
 Art des Abschluß: -

Veranstaltung

einblenden:

Entscheidung

einblenden:

Projektinformation

einblenden:

Quelle

einblenden: