日本語
 
Help Privacy Policy ポリシー/免責事項
  詳細検索ブラウズ

アイテム詳細


公開

会議論文

Interpolated Policy Gradient: Merging On-Policy and Off-Policy Gradient Estimation for Deep Reinforcement Learning

MPS-Authors
/persons/resource/persons217894

Gu,  S.
Dept. Empirical Inference, Max Planck Institute for Intelligent Systems, Max Planck Society;

/persons/resource/persons84193

Schölkopf,  B
Dept. Empirical Inference, Max Planck Institute for Intelligent Systems, Max Planck Society;

External Resource

Link
(全文テキスト(全般))

Fulltext (restricted access)
There are currently no full texts shared for your IP range.
フルテキスト (公開)
公開されているフルテキストはありません
付随資料 (公開)
There is no public supplementary material available
引用

Gu, S., Lillicrap, T., Turner, R. E., Ghahramani, Z., Schölkopf, B., & Levine, S. (2017). Interpolated Policy Gradient: Merging On-Policy and Off-Policy Gradient Estimation for Deep Reinforcement Learning. In I., Guyon, U., von Luxburg, S., Bengio, H., Wallach, R., Fergus, S., Vishwanathan, & R., Garnett (Eds.), Advances in Neural Information Processing Systems 30 (pp. 3849-3858). Curran Associates, Inc. Retrieved from https://papers.nips.cc/paper/6974-interpolated-policy-gradient-merging-on-policy-and-off-policy-gradient-estimation-for-deep-reinforcement-learning.pdf.


引用: https://hdl.handle.net/21.11116/0000-0000-FEA0-D
要旨
要旨はありません